Multi-objective optimization is an essential component of nearly all engineering design. However, for industrial applications, the design process typically demands running expensive computer code and/or real-world experiments putting the design process at risk of finding suboptimal solutions and/or not meeting budget constraints. As a first step toward a remedy, meta-models are built to mimic the response surface at a much lower query cost. We cover a time-tested technology specifically tailored to limited-data scenarios called Bayesian hybrid modeling (GEBHM) developed and maintained at General Electric (GE) research. GEBHM offers Bayesian mean and principled uncertainty predictions allowing a second technology called intelligent design and analysis of experiments (GE-IDACE/IDACE) to perform the optimization task using an adaptive sampling strategy. This chapter first covers the theoretical framework of both GEBHM and GE-IDACE. Then, the impact of GEBHM/GE-IDACE is demonstrated on multiple real-world engineering applications including additive manufacturing, combustion testing, and computational fluid dynamic design modeling. GEBHM and GE-IDACE are used daily and extensively within GE with huge impact in the form of 30–90% cost reduction and superior engineering designs of competitive products.
- intelligent design and analysis of computer experiments
- Bayesian hybrid modeling
- Gaussian process
- adaptive sampling
- surrogate model
- uncertainty sampling
- Bayesian global optimization
- multi-objective optimization
- industrial design
- engineering design
- predictive uncertainty
- machine learning
- artificial intelligence
- robust optimization
- expected improvement
Through better engineering design we can build superior aircraft engines with higher efficiency and lower weight, wind turbines that can produce more energy at a reduced cost, and steam turbines that can reach higher efficiencies, to mention a few examples. The importance of a fast design process which produces global optima with the fewest amount of resources is obviously fundamentally crucial. This process, for the purpose of this chapter, can be thought of as a multi-objective optimization problem. Consider as an example the design of aerodynamic airfoil shapes. First, the performance is produced, followed by a mechanical and aeromechanics assessment. Aeromechanical feedback and reactive aerodynamic redesign rely heavily on the domain expertise of the design engineers. It is not atypical to cycle through 50 of these iterations to obtain a design that satisfies mechanical (stress, creep, fatigue to mention a few) and aeromechanical (say, clean Campbell and flutter resistant) requirements. During these cycles, data from expensive computational codes and/or real-world experiments are collected and the design cycle continues in a direction suggested by this information. Generally speaking, as some examples of the key contributions toward high resource requirements are expensive computer simulations such as computational fluid dynamics (CFD) and ANSYS. In some cases, real-world experiments need to be performed, e.g., when it comes to passing FAA certification.
With this, it should also be clear that engineering design is performed under very strict budgets. Each datum obtained whether from a simulation, physical experiment, or an expert needs to be as informative toward the goals we are trying to accomplish as possible. In some cases, it can take weeks or months to evaluate a single datum. In this case, a meta-model is built on a small representative set of data. This can be Gaussian processes (GPs) [1, 2, 3, 4, 5], Bayesian hybrid modeling as used at GE [6, 7], or polynomial chaos expansions (PCEs) [8, 9].
One of the key goals of this chapter is to present the state-of-the-art industrial tools toward achieving the best possible optima under strict budget constraints. Specifically at GE we are regularly seeing a reduction in cost needed to obtain the same level of information on the order of 30–90%. This leaves more room in the budget for finding even better, more competitive, designs as hitherto possible. As a consequence of the technologies covered in this chapter, we are building better aircraft engines, improving our steam turbines, and harvesting more wind energy because of this. There is still a lot more to be invented and improved, but the following sections will give an idea of where we currently stand.
2. Theoretical framework
In order to introduce some nomenclature and to lay the foundation for surrogate modeling, and adaptive sampling, consider a concrete example of an engineering design task.
Before diving into the details of the setup it is worth briefly discussing how the data is collected. We obtain data in this design task either from real-world experiments or from computer experiments the latter defined in Ref. . A computer experiment consists of running an expensive complex computer code for a set of different inputs. One of the main motivations of using computer codes is to approximate and thereby speed up costly real-world experiments in order to reduce the engineering design cycle time.
Continuing the example, imagine designing a wing blade described by a set of geometric design variables/features/input dimensions. The blade is part of an engine and for a given blade design, i.e., for given values of the features, the engine produces two outputs/objectives one which is the efficiency and the other being the mass flow rate. The design space is an -dimensional space containing the set of all possible geometries we can consider. In this example, we want to maximize the efficiency and to minimize the mass flow rate but in general, we may have objectives. When , we are perform multi-objective optimization, and later in this chapter, we explore a set of sampling strategies suitable for various values of . The set of all points in objective space form a response surface which is assumed to be extremely costly to exhaustively explore putting the search for global optima at risk. Toward reducing the cost of the overall design process regardless of whether the data is obtained by real-world experiments or complex computer codes, a key ingredient is meta-/surrogate modeling where an approximation to the response surface is obtained by generalizing its behavior under certain assumptions (such as smoothness) based on few observed instances of said surface. This is typically done by querying a set of initial points on which the surrogate model is built. Following this, the meta-model is extremely cheap to query in comparison to the surface itself. These meta-models require, typically, only a handful of data to construct an accurate representation of the response surface. An important question however arises on which points to pick, i.e., on how to form the design [10, 12, 13, 14]. Given a budget on the total number of data points, how are new points added to this design sequentially? This is where adaptive sampling comes into play and GEBHM is used in conjunction with a powerful technology called intelligent design and analysis of computer experiments (GE-IDACE/IDACE) [15, 16].
In this section, we provide an overview of the mathematical framework of the GEBHM and GE-IDACE technologies introduced in Section 1. Further theoretical details and application coverage can be found in Refs. [7, 11, 15, 16, 17, 18, 19, 20, 21, 22, 23, 24, 25, 26].
2.1. Bayesian hybrid modeling (GEBHM/BHM)
In industry applications, it is not uncommon for the data to be multi-dimensional, noisy, highly non-linear, and expensive to collect. On a day-to-day basis, we address the challenge of enabling a robust and uncertainty certified design utilizing both limited expansive simulations data and noisy field measurements. GE Research has an in-house software framework for advanced Bayesian modeling and machine learning named GEBHM, sometimes we shall refer to this as simply BHM, which enables the combination of multiple numerical simulations and experimental sources of data in one unified workflow. As shown in Figure 1, GEBHM capabilities are: uncertainty propagation and quantification, sensitivity analysis, full Bayesian model calibration, meta-modeling, multi-fidelity analysis, and adaptive design of numerical experiments. The theoretical framework of the GEBHM is based on Kennedy O’Hagans approach to modeling and fusing simulation and experimental data with associated uncertainties using GPs . The noisy high-fidelity model is represented as Gaussian process aggregated from a linear combination of a low-fidelity model and a model discrepancy function as
where is the (high-fidelity numerical model or experimentally measured) response. The low-fidelity model is and discrepancy term are modeled by separate GPs. The design variable is denoted by , while the calibration parameters are denoted by . GEBHM allows for calibration of a set of model parameters in the low-fidelity model. For example, this could be parameters in a CFD simulation that we want to tune/calibrate in order to match real-world experimental runs as closely as possible. The measurement error is denoted by .
The GP hyperparameters are learned using a Markov Chain Monte Carlo (MCMC) technique based on an Metropolis-Hastings-within-Gibbs algorithm [27, 28] with univariate proposal distributions for the posterior distribution updates. MCMC generally converges toward the most probable values for the parameters which best explain the data  from which representative samples can be obtained. To avoid overfitting the high-fidelity data, the initial values of the hyperparameters of covariance matrices are updated with the current realizations at every MCMC step, and realizations from the posterior distributions of the model parameters are produced.
2.2. Intelligent design and analysis of computer experiments (GE-IDACE/IDACE)
While meta-models offer very low query times for response surface values they are still only approximate, especially when built on small datasets. Thus, having the meta-model does not generally mean that we can use this entirely in place of the response surface. However, we can use it as a guide to seek out new locations in the design space that are promising toward our goal which could, e.g., be optimization or to produce the most accurate surrogate model possible. The focus in this chapter is on the prior goal of global optimization.
GE-IDACE, also sometimes referred to as simply IDACE, uses the expected improvement (EI)  method to explore and exploit the design space for obtaining the global optimum with the fewest possible resources, see Figure 2. Without loss of generality, EI defines the improvement of a new design point as , where is the current best point (also called the incumbent) and we will suppress going forward. The surrogate model predicts a distribution for denoted . This makes a random variable. In face of this randomness, we are just interested in knowing how large the improvement is expected to be:
In each iteration of GE-IDACE, the point with maximum EI is added to the design.
In what follows, we review some GE-IDACE multi-objective optimization EI methods that we have found to work well in practice but emphasize that more research is needed toward getting faster at locating the global optimum in multiple dimensions under increasingly stricter budgets .
2.2.1. Multiple objectives: centroid method for two dimensions
Many approaches exist for running IDACE with multiple objectives [31, 32, 33, 34, 35, 36, 37, 38, 39]. Here, consider the two-dimensional case and the so-called centroid method which shares a similar intuition as its one-dimensional counterpart Eq. (3). In this methodology, each candidate point from the design space is imagined to create a centroid point, the equations for computing this point are given below. This centroid point, which is located in output space, is then compared to its closest Pareto point on the current frontier [40, 41]. For simplicity, consider now two different candidate/design points where we compute the associated centroid point for each. Then, for the th candidate point the centroid point in two-dimensional output space is , the Pareto point on the current frontier closest to is computed and the distance between and quantifies the value of adding candidate point . Note that in general. The candidate point inducing the biggest expected change in the Pareto frontier as measured by the distance between its centroid point and corresponding Pareto point is picked in a given iteration.
The probability that a new design point at improves the th member of the current Pareto front, denoted , is:
The two-dimensional EI then becomes:
where is the Pareto point, among all Pareto points of the current Pareto frontier, which is closest to the centroid point . The details are available in Ref.  where also another method assuming a convex hull Pareto shape is discussed, which in certain scenarios outperforms the centroid method.
2.2.2. Multiple objectives: hypervolume method for any dimensions
The hypervolume EI method is presented to handle high-dimensional objective spaces beyond two [15, 47]. Drawing an analogy with the one-dimensional case, in multi-dimensional objective space, the hypervolume is considered a measure of the current known minimum point (the Pareto front). The difference in the hypervolume between the current Pareto front and the new Pareto front resulting from adding a candidate point is used to define the EI. Accordingly, the EI in multi-dimensional objective-space, , gained by adding a new point with objective values to the design space is defined as
where denotes the hypervolume contained by the current Pareto front. In general, the expectation integral in the hypervolume EI method is simplified by decomposing it into sub integrals over hyperrectangles, please see Ref. . Further simplification to this integral can be achieved by assuming the predicted output components are independent . To reduce the cost of computing the hyperrectangles integrals, Monte Carlo approximation can be utilized  as well as using a suitable merging approach to decompose the integral as presented in . It is worth mentioning that depending on the domination of the proposed point over the current Pareto front, different levels of improvement can be gained .
2.2.3. GE-IDACE with desirability
The physical programming technique can help the engineer to guide the design algorithm toward the desirable regions of the Pareto front . For example, let the th objective in a multi-objective design problem be denoted by . Then, by dividing the range of possible values into four segments (e.g., , , , and ), the design engineer can assign a desirability for each sub-range as a highly desirable, acceptable, undesirable, and unacceptable. An aggregate desirability function that is formed from these individual ranges is used to rank the Pareto points. Within the GE-IDACE framework, we help the designer to define desirabilities from a set of functions by decomposing the objectives into ranges as shown in Figure 3. The desirability function in this case is one-dimensional and decomposed into three regions whereby different desirability functions are assigned by the designer. The figure shows that the candidate points mostly favored have an objective value predicted to be in the range (−4, 4).
Next, we extend the hyperrectangle approach to account for desirability as follows. Specifically, the designer chooses, for each objective, ranges of the objective which are considered highly desirable, acceptable, undesirable, and unacceptable. Closely following the ideas in Ref. , consider the following quantity called expected desirability of improvement (EDI):
Given the predictive distribution for some new point, is the mean desirability of the predictions that improve the current Pareto front.
3. Industrial applications of GEBHM/GE-IDACE
Having covered the theoretical framework for both GEBHM and GE-IDACE this section turns to demonstrate their application to real-world engineering applications. We consider first additive manufacturing (AM) which is a vital process for many engineering design applications and is bound to further transform manufacturing. In essence, AM can be defined as the process of overlying layers to create a three-dimensional objects [47, 48]. We will show how GE-IDACE reduces the design cycle time from 6 months to a few weeks.
As a second application, we consider combustion testing where the goal in this case is to maximize load while keeping exhaust and temperature within specific limits. We demonstrate that GE-IDACE can help guide the test into regions with 20% more points from critical areas compared to status quo.
Finally, we demonstrate how well GE-IDACE does for expensive complex computer codes such as CFD modeling. We show that GEBHM/GE-IDACE helps reduce the number of test points by a factor of three when compared with neural network modeling and optimization.
3.1. Additive manufacturing
As a main example of AM applications utilizing GE-IDACE we consider Direct Metal Laser Melting (DMLM) but mention that GE-IDACE is also used for feature-based qualification methods for Directed Energy Deposition having a big positive impact.
DMLM is a key modality of additive manufacturing that focuses on 3D printing of metallic materials. Printing metals is in itself a complicated task due to the microstructural instabilities from melting of the metallic powder. It is especially complicated for superalloys since as-built parts from DMLM are highly prone to microcracking and other microstructural deficiencies. So it is of primary importance to identify what the processing parameters are for the hard-to-process Nickel-based superalloys, and that process has been proven to be non-trivial. The lead times for processing parameter development for these types of alloys are typically on the order of several weeks to months, which means increased cost and the inability to introduce new materials in the additive marketplace. In order to reduce the cycle time when developing the processing parameters for DMLM for hard-to-process alloys, we have extensively utilized GEBHM and GE-IDACE to collect data in an intelligent manner . Typically the key parameters that dictate the processing of additive parts are the laser power, the laser speed, etc. GE-IDACE automated the process toward obtaining design points, i.e., processing parameter combinations, which were most informative to the model and which would guide us in the direction of the optimal solution(s). We used quantified characteristics of the microstructural deficiencies as our outputs/objectives. Parts were built in the additive machine and then characterized by sectioning the parts, imaging the sections, and performing automated image analysis. This enabled us to analyze a large number of images extracting specific defect information such as porosity, keyholes due to unmelted powder, etc. We are interested in porosity because it affects the mechanical properties like yield strength and fatigue life adversely. The GE-IDACE process then constructed a model of the output microstructural defects as function of the input process parameters. We utilized both variance minimization (uncertainty sampling) and EI-based optimization to exploit and explore the design space to identify the optimal solutions faster. Using GE-IDACE with GEBHM, we were able to reduce the cycle time in identifying the optimal process parameter window for a superalloy from 6 months to a few weeks.
Figure 4 shows the progression of the collection of data (only two dimensions shown in a multi-dimensional problem). We can clearly see that the GE-IDACE methodology helps us explore the design space initially and then start to exploit the optimal solutions in the later iterations to quickly converge on an optimal processing window. This shows that by the third iterations GE-IDACE suggested to us almost 65% of points that satisfy our objectives in defects, while also bringing the overall model uncertainty down. Currently, we are working on expanding this methodology into more complicated structures and additional quantities of interest (QoIs) such as mechanical properties, durability, surface finish etc.
3.2. Combustion testing
During manufacturing of turbine or jet engines, combustion testing is required at different stages of development, manufacturing, installation, and deployment to ensure that the engine is working as designed and within desired tolerances. Multiple of such experiments are required for different operating points making this process very time consuming and expensive. Traditionally, test plans are created prior to actual experiments which consists of heuristics test blocks or groups of tests. These test blocks are generally created from expert judgment based on prior operational experience. However, these test plans may not be optimal as they are heuristically designed without rigorous statistical analysis of legacy and existing data. The obvious results is that traditional heuristic test plans may lead to inefficient and redundant allocation of resources.
Therefore, GE-IDACE can be employed to improve the test schedule. This happens by adaptively learning the system characteristics and performance with the underlying advanced surrogate model GEBHM as the function of input conditions using real-time data or even leverage historical experiments while also incorporating expert judgment. The test plan is hereby dynamically learned, compiled, ranked, and updated.
Ideally, historical data is available. The first step is to build GEBHM on this dataset. The input variables in this application are gas splits, loads, speed, firing temperature, etc., and the outputs are NOx emission, combustor instability, system dynamics, etc. The process followed with GE-IDACE is as shown in Figure 2. The steps are repeated as more data is added until necessary goals in the form of certification approval, optimum operating conditions, and/or constraints in the shape of time and budget are all met as an example.
The user typically possess desirability ratings for experimental outcomes. This desirability may include a factor, threshold, constraint, goal, or objective that is important to the user for testing such as emissions thresholds, maximum and minimum loads, efficiency ratings, among others. For example, the user would like to know the operating conditions at which the load is maximum while NOx exhaust and temperature are within some limits. The desirability is provided by the user as target values, target ranges, or by a custom function over the quantity of interest; please see Section 2.2.3 for a review.
With this introduction, in the following we demonstrate the impact of using GE-IDACE on combustion testing. A design space of four operating conditions , and are explored, such that two performance parameters and stay within some thresholds defined as: and . The goal is to design a test plan to maximize the number of experiments within said thresholds.
First, for later comparison, the traditional approach with one-factor-at-a-time designs are shown in Figure 5. Grey points indicate experiments out-of-bounds from a threshold perspective. Blue points met the conditions, i.e., they are within the blue delineated region of objective space. Out of a total of 69 experiments performed, 10 (14.5%) satisfied the desirability of , 27 (39.1%) satisfied the desirability of , and 35 (50.7%) satisfied the desirability of both and .
Then, as an aim to improve this process, GE-IDACE was used to carry out a dynamic test plan. After each experiment, GEBHM was updated on the new data and the next point was picked based on the desirability with regards to the output responses. The corresponding output performance of these experiments and desirable regions is shown in Figure 5B to be compared with Figure 5A. Out of a total of 69 experiment performed, 25 (36.2%) satisfied the desirability of , 40 (57.9%) satisfied the desirability of , and 47 (68.2%) satisfied the desirability of both and . The impact is that GE-IDACE increases the number of points in the desirable region by 20% with the same number of tests. Given the high cost of running these experiments, this easily translates to hundreds of thousands of dollars saved annually.
In summary, the impact of GE-IDACE is clear to combustion testing. In fact, the application has now extending beyond just combustion testing: we are performing full-scale engine testing as well with this new strategy and it is being rolled out to multiple GE’s businesses thus achieving severe cost reductions and better engineering designs.
3.3. Computational fluid dynamics for turbine design
Next, we consider the application of GE-IDACE to turbine design with CFD. So far in this chapter, we have covered real-world engineering applications. In the following, we demonstrate how GE-IDACE is positively impacting expensive computer simulations as well.
Aerodynamic optimization of a turbine involves dozens of variables, impacting everything from system level features through detailed airfoil properties. Two primary top-level considerations for the aerodynamic design of a turbine include vortexing and airfoil stack. Vortexing involves custom tailoring of the vane and rotor exit angle distributions. This establishes the radial distribution of work within the turbine stage. Vortexing affects local acceleration and mass flow distributions, and thus is a strong driver of secondary loss generation (endwall vortices). Airfoil stacking aerodynamically imposes body forces on the flow, further affecting the radial mass flow and work distributions. Stacking also strongly influences the generation of secondary loss. The general objective of a vortexing and stack optimization is to maximize turbine performance, usually through management of secondary loss growth, while also adhering to numerous constraints that ensure proper downstream performance and acceptable component life.
Before covering how GE-IDACE improved the optimization, consider the traditional approach as shown in Figure 6. The first stage vane is optimized using a component-specific space-filling DOE on which CFD is evaluated. These results are then used to build a surrogate model that characterizes a row-specific loss metric (relative total pressure loss or secondary kinetic energy, for example). A genetic algorithm (GA) is used to optimize a set of X’s (defining a design point) describing the geometry for minimal loss (maximum efficiency) based on the surrogate model. This process is repeated for each subsequent row, with downstream components reacting to the results of the upstream row’s optimized exit flow conditions.
Compared to the traditional approach, GE-IDACE can help automate the geometry generation process to ensure efficient throughput. To accelerate build time from an X-specification to a CFD-ready geometry, a mesh morphing approach is implemented. Leveraging the block structured hexahedral mesh from the baseline geometry’s CFD analysis, new cases re-stretch the inlet, exit, and passage blocks to produce a topologically identical mesh that conforms to the new 3D airfoil surface. The O block surrounding airfoil remains largely unchanged and translates with the new geometry. The baseline mesh is similar in fidelity to a typical “production” CFD analysis for turbine design. Surface y + is ∼1 for all airfoil metal surfaces, and in total, the high pressure turbine (HPT) domain consists of ∼9 million nodes. Figure 7 shows a representative example of the baseline grid and how it is morphed to an updated geometry.
All processes required to translate X’s to CFD geometries are batch enabled, and each new CFD case requires ∼15 min of wall clock time to generate. The CFD analysis is performed using GE’s in-house CFD solver, TACOMA. TACOMA is a 2nd order accurate (in time and space), finite-volume, block-structured, compressible flow solver, implemented in Fortran 90. Stability is achieved via the JST scheme, and convergence is accelerated using pseudo-time marching and multi-grid techniques. The Reynolds Averaged Navier-Stokes (RANS) equations are closed via the k-turbulence model of Wilcox. Multi-row analysis is enabled through the use of mixing plane interfaces. Using 64 total CPU cores for the four-airfoil HPT domain, convergence is achieved in roughly 6 h.
The objective of the blade design task will be group efficiency. This metric is evaluated for each candidate point as a delta from a known baseline, which for this case is a modern two-stage Aviation HPT that already leverages results from prior optimization using the traditional techniques described earlier. All four HPT airfoils are considered in this optimization. To establish an entitlement performance, no constraints are imposed at this time to account for mechanical requirements or downstream component performance. Traditional space-filling DOEs for high dimensional problems require a large number of data points, and for a CFD-based study, an out-of-budget amount of computational resources. To manage these requirements, and to maintain design-cycle-relevant optimization times, advanced machine learning techniques are employed to intelligently guide the optimization process.
As a benchmark for the new process a 320-point space-filling OLH DOE was first created to cover all 32 HPT variables. Consistent with current practice, a radial basis function (RBF) surrogate model was fit to this data set, and GA optimization was performed on the RBF model. Modest gains over baseline were achieved from this approach.
For GE-IDACE, a more sparsely populated OLH DOE of 50 points was generated to seed the optimization. Leveraging GEBHM capabilities, additional DOE points were added only in areas of high error where the GEBHM model predicted high likelihood of progressing toward the objective—maximum delta group efficiency over the baseline. Through several rounds of intelligent incremental point addition, where each round included a refinement to the GEBHM fit, a new final optimal was established that exceeded the previous delta by roughly three times. Additionally, as shown in Figure 8, this much more favorable outcome was achieved with roughly a third of the computational resources.
4. Summary and future work
It has been demonstrated how advanced engineering tools centered around adaptive sampling in multi-objective space help achieve better engineering designs at highly-reduced cost. The underlying technologies are GEBHM and GE-IDACE which were covered first from a theoretical perspective. Then, applications in the areas of additive manufacturing, combustion testing, and computational fluid dynamics were considered. The impact of using GEBHM/GE-IDACE was clear and far surpassed status quo. At GE we consistently find a 30–90% resource cost reduction.
Before discussing future work, we first cover some of the main limitations of the GE-IDACE tool. Fundamentally, GE-IDACE treats the computer experiment as a black box function, i.e., it only sees inputs to the code and the corresponding outputs. In some cases, this information is all we are able to leverage, but in other situations we may have additional insights which, if taken advantage of, could speed up the optimization. For example, gradient information could be available from the experiments too. Furthermore, the GE-IDACE approach is “greedy,” i.e., it selects the next input point from the design space which is predicted to give the best immediate outcome with the current state of knowledge. This approach might not be the optimum strategy in the long-term. Worded differently, under a budget, there could exist a possibility that one can reach a better overall solution with fewer experiments without selecting the highest EI point in each intermediate step. Finally, it is difficult to thoroughly parallelize experiments with GE-IDACE as it is a sequential process which requires data acquisition and model-updating as the experimental results are available, although some approximate schemes exist .
In terms of future work and further improvements, we demonstrate in Ref.  that Particle Swarm Optimization performs very well for EI computation. A lot of exciting opportunities exist for GEBHM and GE-IDACE to further improve the engineering design process and remain to be discovered. In recent work, we demonstrate how to use GE-IDACE with multi-fidelity data sources (simulation vs. experiments, e.g.,)  and how to leverage legacy data from other designs into the GEBHM modeling process , to reduce the cost of running tests for new engine designs. In terms of future work, GEBHM can be extended to operate fluently across any type of data in terms of dimensionality and number of points. This way, all the benefits of GEBHM and GE-IDACE can be leveraged at any scale. Toward this, an initial exploration of a parallelizable way to fit the GEBHM is found in Ref. . This extends the size of datasets which GEBHM can fit by a factor of 5–10.
The authors acknowledge support from GE Research and are grateful to Tom Vandeputte at GE Research for details on the CFD modeling.