Jump to content
 







Main menu
   


Navigation  



Main page
Contents
Current events
Random article
About Wikipedia
Contact us
Donate
 




Contribute  



Help
Learn to edit
Community portal
Recent changes
Upload file
 








Search  

































Create account

Log in
 









Create account
 Log in
 




Pages for logged out editors learn more  



Contributions
Talk
 



















Contents

   



(Top)
 


1 Introduction  



1.1  Linear speedup  





1.2  Noise  







2 Approach  





3 Predictor equations  



3.1  Kriging  





3.2  Indirect GEK  





3.3  Direct GEK (through prior covariance matrix)  





3.4  Direct GEK (through observation matrix)  





3.5  Gradient-enhanced kriging for high-dimensional problems (Indirect method)  





3.6  Augmented gradient-enhanced kriging (direct method)  







4 Example: Drag coefficient of a transonic airfoil  





5 Applications  





6 References  














Gradient-enhanced kriging







Add links
 









Article
Talk
 

















Read
Edit
View history
 








Tools
   


Actions  



Read
Edit
View history
 




General  



What links here
Related changes
Upload file
Special pages
Permanent link
Page information
Cite this page
Get shortened URL
Download QR code
Wikidata item
 




Print/export  



Download as PDF
Printable version
 
















Appearance
   

 






From Wikipedia, the free encyclopedia
 


Gradient-enhanced kriging (GEK) is a surrogate modeling technique used in engineering. A surrogate model (alternatively known as a metamodel, response surface or emulator) is a prediction of the output of an expensive computer code.[1] This prediction is based on a small number of evaluations of the expensive computer code.

Introduction

[edit]
Example of one-dimensional data interpolated by Kriging and GEK. The black line indicates the test-function, while the gray circles indicate 'observations', 'samples' or 'evaluations' of the test-function. The blue line is the Kriging mean, the shaded blue area illustrates the Kriging standard deviation. With GEK we can add the gradient information, illustrated in red, which increases the accuracy of the prediction.

Adjoint solvers are now becoming available in a range of computational fluid dynamics (CFD) solvers, such as Fluent, OpenFOAM, SU2 and US3D. Originally developed for optimization, adjoint solvers are now finding more and more use in uncertainty quantification.

Linear speedup

[edit]

An adjoint solver allows one to compute the gradient of the quantity of interest with respect to all design parameters at the cost of one additional solve. This, potentially, leads to a linear speedup: the computational cost of constructing an accurate surrogate decrease, and the resulting computational speedup scales linearly with the number of design parameters.

The reasoning behind this linear speedup is straightforward. Assume we run primal solves and adjoint solves, at a total cost of . This results in data; values for the quantity of interest and partial derivatives in each of the gradients. Now assume that each partial derivative provides as much information for our surrogate as a single primal solve. Then, the total cost of getting the same amount of information from primal solves only is . The speedup is the ratio of these costs: [2] [3]

A linear speedup has been demonstrated for a fluid-structure interaction problem [2] and for a transonic airfoil.[3]

Noise

[edit]

One issue with adjoint-based gradients in CFD is that they can be particularly noisy. [4] [5] When derived in a Bayesian framework, GEK allows one to incorporate not only the gradient information, but also the uncertainty in that gradient information.[6]

Approach

[edit]

When using GEK one takes the following steps:

  1. Create a design of experiment (DoE): The DoE or 'sampling plan' is a list of different locations in the design space. The DoE indicates which combinations of parameters one will use to sample the computer simulation. With Kriging and GEK, a common choice is to use a Latin Hypercube Design (LHS) design with a 'maximin' criterion. The LHS-design is available in scripting codes like MATLABorPython.
  2. Make observations: For each sample in our DoE one runs the computer simulation to obtain the Quantity of Interest (QoI).
  3. Construct the surrogate: One uses the GEK predictor equations to construct the surrogate conditional on the obtained observations.

Once the surrogate has been constructed it can be used in different ways, for example for surrogate-based uncertainty quantification (UQ) or optimization.

Predictor equations

[edit]

In a Bayesian framework, we use Bayes' Theorem to predict the Kriging mean and covariance conditional on the observations. When using GEK, the observations are usually the results of a number of computer simulations. GEK can be interpreted as a form of Gaussian process regression.

Kriging

[edit]

Along the lines of, [7] we are interested in the output of our computer simulation, for which we assume the normal prior probability distribution:

with prior mean and prior covariance matrix . The observations have the normal likelihood:

with the observation matrix and the observation error covariance matrix, which contains the observation uncertainties. After applying Bayes' Theorem we obtain a normally distributed posterior probability distribution, with Kriging mean:

and Kriging covariance:

where we have the gain matrix:

In Kriging, the prior covariance matrix is generated from a covariance function. One example of a covariance function is the Gaussian covariance:

where we sum over the dimensions and are the input parameters. The hyperparameters , and can be estimated from a Maximum Likelihood Estimate (MLE).[6] [8] [9]

Indirect GEK

[edit]

There are several ways of implementing GEK. The first method, indirect GEK, defines a small but finite stepsize , and uses the gradient information to append synthetic data to the observations , see for example.[8] Indirect Kriging is sensitive to the choice of the step-size and cannot include observation uncertainties.

Direct GEK (through prior covariance matrix)

[edit]

Direct GEK is a form of co-Kriging, where we add the gradient information as co-variables. This can be done by modifying the prior covariance or by modifying the observation matrix ; both approaches lead to the same GEK predictor. When we construct direct GEK through the prior covariance matrix, we append the partial derivatives to , and modify the prior covariance matrix such that it also contains the derivatives (and second derivatives) of the covariance function, see for example [10] .[6] The main advantages of direct GEK over indirect GEK are: 1) we do not have to choose a step-size, 2) we can include observation uncertainties for the gradients in , and 3) it is less susceptible to poor conditioning of the gain matrix . [6] [8]

Direct GEK (through observation matrix)

[edit]

Another way of arriving at the same direct GEK predictor is to append the partial derivatives to the observations and include partial derivative operators in the observation matrix , see for example.[11]

Gradient-enhanced kriging for high-dimensional problems (Indirect method)

[edit]

Current gradient-enhanced kriging methods do not scale well with the number of sampling points due to the rapid growth in the size of the correlation matrix, where new information is added for each sampling point in each direction of the design space. Furthermore, they do not scale well with the number of independent variables due to the increase in the number of hyperparameters that needs to be estimated. To address this issue, a new gradient-enhanced surrogate model approach that drastically reduced the number of hyperparameters through the use of the partial-least squares method that maintains accuracy is developed. In addition, this method is able to control the size of the correlation matrix by adding only relevant points defined through the information provided by the partial-least squares method. For more details, see.[12] This approach is implemented into the Surrogate Modeling Toolbox (SMT) in Python (https://github.com/SMTorg/SMT), and it runs on Linux, macOS, and Windows. SMT is distributed under the New BSD license.

Augmented gradient-enhanced kriging (direct method)

[edit]

A universal augmented framework is proposed in [9] to append derivatives of any order to the observations. This method can be viewed as a generalization of Direct GEK that takes into account higher-order derivatives. Also, the observations and derivatives are not required to be measured at the same location under this framework.

Example: Drag coefficient of a transonic airfoil

[edit]
Transonic airfoil.
Reference results for the drag coefficient of a transonic airfoil, based on a large number of CFD simulations. The horizontal and vertical axis show the deformation of the shape of the airfoil.
Kriging surrogate model of the drag coefficient of a transonic airfoil. The gray dots indicate the configurations for which the CFD solver was run.
GEK surrogate model of the drag coefficient of a transonic airfoil. The gray dots indicate the configurations for which the CFD solver was run, the arrows indicate the gradients.

As an example, consider the flow over a transonic airfoil.[3] The airfoil is operating at a Mach number of 0.8 and an angle of attack of 1.25 degrees. We assume that the shape of the airfoil is uncertain; the top and the bottom of the airfoil might have shifted up or down due to manufacturing tolerances. In other words, the shape of the airfoil that we are using might be slightly different from the airfoil that we designed.

On the right we see the reference results for the drag coefficient of the airfoil, based on a large number of CFD simulations. Note that the lowest drag, which corresponds to 'optimal' performance, is close to the undeformed 'baseline' design of the airfoil at (0,0).

After designing a sampling plan (indicated by the gray dots) and running the CFD solver at those sample locations, we obtain the Kriging surrogate model. The Kriging surrogate is close to the reference, but perhaps not as close as we would desire.

In the last figure, we have improved the accuracy of this surrogate model by including the adjoint-based gradient information, indicated by the arrows, and applying GEK.

Applications

[edit]

GEK has found the following applications:

References

[edit]
  1. ^ Mitchell, M.; Morris, M. (1992). "Bayesian design and analysis of computer experiments: two examples" (PDF). Statistica Sinica (2): 359–379.
  • ^ a b c de Baar, J.H.S.; Scholcz, T.P.; Verhoosel, C.V.; Dwight, R.P.; van Zuijlen, A.H.; Bijl, H. (2012). "Efficient uncertainty quantification with gradient-enhanced Kriging: Applications in FSI" (PDF). ECCOMAS, Vienna, Austria, September 10–14.
  • ^ a b c d de Baar, J.H.S.; Scholcz, T.P.; Dwight, R.P. (2015). "Exploiting Adjoint Derivatives in High-Dimensional Metamodels". AIAA Journal. 53 (5): 1391–1395. Bibcode:2015AIAAJ..53.1391D. doi:10.2514/1.J053678.
  • ^ Dwight, R.; Brezillon, J. (2006). "Effect of Approximations of the Discrete Adjoint on Gradient-Based Optimization". AIAA Journal. 44 (12): 3022–3031. Bibcode:2006AIAAJ..44.3022D. CiteSeerX 10.1.1.711.4761. doi:10.2514/1.21744.
  • ^ Giles, M.; Duta, M.; Muller, J.; Pierce, N. (2003). "Algorithm Developments for Discrete Adjoint Methods". AIAA Journal. 41 (2): 198–205. Bibcode:2003AIAAJ..41..198G. doi:10.2514/2.1961. S2CID 2106397.
  • ^ a b c d e de Baar, J.H.S.; Dwight, R.P.; Bijl, H. (2014). "Improvements to gradient-enhanced Kriging using a Bayesian interpretation". International Journal for Uncertainty Quantification. 4 (3): 205–223. doi:10.1615/Int.J.UncertaintyQuantification.2013006809.
  • ^ Wikle, C.K.; Berliner, L.M. (2007). "A Bayesian tutorial for data assimilation". Physica D. 230 (1–2): 1–16. Bibcode:2007PhyD..230....1W. doi:10.1016/j.physd.2006.09.017.
  • ^ a b c d Dwight, R.P.; Han, Z.-H. (2009). Efficient uncertainty quantification using gradient-enhanced Kriging (PDF). doi:10.2514/6.2009-2276. ISBN 978-1-60086-975-4. S2CID 59019628. {{cite book}}: |journal= ignored (help)
  • ^ a b Zhang, Sheng; Yang, Xiu; Tindel, Samy; Lin, Guang (2021). "Augmented Gaussian random field: Theory and computation". Discrete & Continuous Dynamical Systems - S. 15 (4): 931. arXiv:2009.01961. doi:10.3934/dcdss.2021098. S2CID 221507566.
  • ^ a b Laurenceau, J.; Sagaut, P. (2008). "Building efficient response surfaces of aerodynamic functions with Kriging and coKriging". AIAA Journal. 46 (2): 498–507. Bibcode:2008AIAAJ..46..498L. doi:10.2514/1.32308. S2CID 17895486.
  • ^ de Baar, J.H.S. (2014). "Stochastic Surrogates for Measurements and Computer Models of Fluids". PhD Thesis, Delft University of Technology: 99–101.
  • ^ Bouhlel, M.A.; Martins, J.R.R.A. (2018). "Gradient-enhanced kriging for high-dimensional problems". Engineering with Computers. 35: 157–173. arXiv:1708.02663. doi:10.1007/s00366-018-0590-x. S2CID 3540630.
  • ^ Morris, M.D.; Mitchell, T.J.; Ylvisaker, D. (1993). "Bayesian Design and Analysis of Computer Experiments: Use of Derivatives in Surface Prediction". Technometrics. 35 (3): 243–255. doi:10.1080/00401706.1993.10485320.
  • ^ Chung, H.-S.; Alonso, J.J. (2002). "Using Gradients to Construct Cokriging Approximation Models for High-Dimensional Design Optimization Problems". AIAA 40th Aerospace Sciences Meeting and Exhibit: 2002–0317. CiteSeerX 10.1.1.12.4149. doi:10.2514/6.2002-317.
  • ^ Han, Z.-H.; Gortz, S.; Zimmermann, R. (2013). "Improving variable-fidelity surrogate modeling via gradient-enhanced kriging and a generalized hybrid bridge function". Engineering with Computers. 32 (1): 15–34. doi:10.1016/j.ast.2012.01.006.
  • ^ Ulaganathan, S.; Couckuyt, I.; Dhaene, T.; Degroote, J.; Laermans, E. (2016). "Performance study of gradient-enhanced Kriging". Aerospace Science and Technology. 25 (1): 177–189.
  • ^ Laurent, L.; Le Riche, R.; Soulier, B.; Boucard, P.-A. (2017). "An overview of gradient-enhanced metamodels with applications" (PDF). Archives of Computational Methods in Engineering. 26: 1–46. doi:10.1007/s11831-017-9226-3. S2CID 54625655.
  • ^ Lockwood, B.A.; Anitescu, M. (2012). "Gradient-Enhanced Universal Kriging for Uncertainty Propagation" (PDF). Nuclear Science and Engineering. 170 (2): 168–195. CiteSeerX 10.1.1.187.6097. doi:10.13182/NSE10-86. S2CID 18465024.
  • ^ Raggi, G.; Fdez. Galván, I.; Ritterhoff, C. L.; Vacher, M.; Lindh, R. (2020). "Restricted-Variance Molecular Geometry Optimization Based on Gradient-Enhanced Kriging". Journal of Chemical Theory and Computation. 16 (6): 3989–4001. doi:10.1021/acs.jctc.0c00257. PMC 7304864. PMID 32374164.

  • Retrieved from "https://en.wikipedia.org/w/index.php?title=Gradient-enhanced_kriging&oldid=1136094389"

    Categories: 
    Mathematical modeling
    Computational fluid dynamics
    Hidden categories: 
    CS1 errors: periodical ignored
    Articles with short description
    Short description matches Wikidata
    Wikipedia articles with possible conflicts of interest from April 2017
    Articles that may contain original research from April 2017
    All articles that may contain original research
     



    This page was last edited on 28 January 2023, at 18:33 (UTC).

    Text is available under the Creative Commons Attribution-ShareAlike License 4.0; additional terms may apply. By using this site, you agree to the Terms of Use and Privacy Policy. Wikipedia® is a registered trademark of the Wikimedia Foundation, Inc., a non-profit organization.



    Privacy policy

    About Wikipedia

    Disclaimers

    Contact Wikipedia

    Code of Conduct

    Developers

    Statistics

    Cookie statement

    Mobile view



    Wikimedia Foundation
    Powered by MediaWiki