Elsevier

Knowledge-Based Systems

Volume 184, 15 November 2019, 104903
Knowledge-Based Systems

Gradient-enhanced high dimensional model representation via Bayesian inference

https://doi.org/10.1016/j.knosys.2019.104903Get rights and content

Highlights

  • HDMR surrogate model are developed based on Bayesian inference technique.

  • An efficient method for HDMR model integrating gradient information is presented.

  • Eight benchmark examples are used to validate the effectiveness of the method.

Abstract

Recently, gradient-enhanced surrogate models have drawn extensive attention for function approximation, in which the gradient information is utilized to improve the surrogate model accuracy. In this work, gradient-enhanced high dimensional model representation (HDMR) is established based on Bayesian inference technique. The proposed method first assigns Gaussian process prior for the model response function and its partial derivative functions (with respect to all the input variables). Then the auto-covariance functions and the cross-covariance functions of these random processes are established respectively by the HDMR basis functions. Finally, the posterior distribution of the response function is analytically obtained through Bayes theorem. The proposed method combines the sample information and gradient information in a seamless way to yield a highly accurate HDMR prediction model. We demonstrate our method via several examples, and the results all suggest that combining gradient information with sample information provides more accurate prediction results at reduced computational cost.

Introduction

Over the last decades, along with the rapid development of computer science and technique, a variety of complex computational models have been developed for simulating and predicting the behavior of systems in nearly all fields of engineering and science. In the meanwhile, surrogate model is becoming more and more popular to meet the growing engineering demand in computational power [1]. For decades, various types of surrogate model techniques have been continuously developed, such as Kriging/Gaussian Process (GP) [2], [3], [4], [5], [6] and its gradient-enhanced versions [7], support vector regression (SVR) [8], [9], [10], [11], [12]and its gradient-enhanced variant [13], [14], [15], radial basis functions (RBF) [16], [17], [18], polynomial chaos expansion (PCE) [13], [19], [20], [21], [22], [23], artificial neural networks (ANN) [24], [25], [26] and high dimensional model representation (HDMR) [27], [28], [29], [30], [31], [32], [33], [34], [35], among which the HDMR has attracted significant interest due to its ability to handle complex high dimensional models.

HDMR is developed as a set of quantitative model assessment and analysis tools for capturing complex high-dimensional input–output system behavior [31], [32], [36]. Its theoretical foundation lies in that only low-order correlations amongst the input variables have a significant effect on the model outputs, thus the HDMR permits expressing a multidimensional function as a sum of many low dimensional component functions to reflect the independent and cooperative contribution. In practical applications, two types of HDMR expansions are developed in literature. One decomposes models output into components defined on hyper-planes based on a fixed cut point or anchor point in the parameter space, thus this method is known as Cut-HDMR. The other method conforms to the analysis of variance (ANOVA) decomposition widely used in uncertainty quantification and sensitivity analysis, and hence is called ANOVA-HDMR.

To construct the HDMR surrogate model, various methods are developed to evaluate the low-order component terms of the truncated dimension-wise decompositions. For Cut-HDMR, the usual practice is to set up a look-up table of the component functions values evaluated at the selected cut points or anchor points in the parameter space, and the Lagrange or linear interpolation method is utilized to estimate the component functions. Rabitz et al. [37] used equidistantly distributed sample points along each axis of the input parameter to develop Cut-HDMR. Liu et al. [27] suggested using non-uniform optimal nodes to mitigate the Runge’s phenomenon. The optimal nodes are selected as the nodes of Legendre–Gauss, Chebyshev—Gauss and Clenshaw–Curtis quadratures. ANOVA-HDMR, also known as random sampling (RS)-HDMR, expands the component functions in terms of analytical basis functions, and the Monte Carlo simulations or other technique is utilized to estimate the basis function coefficients. Li et al. [36] suggested using orthogonal polynomials, cubic B-splines and polynomial to approximate the ANOVA-HDMR component functions. Luo et al. [34] used the reproducing kernel technique to estimate the arbitrary order ANOVA-HDMR component functions. Lambert et al. [38] utilized the group method of data handling (GMDH) method to construct the ANOVA-HDMR surrogate model based on Legendre polynomial basis functions.

This work aims at developing the ANOVA-HDMR surrogate model integrating with gradient information based on Bayesian inference technique, namely, gradient-enhanced HDMR (GE-HDMR). Firstly, the response function of an underlying model is approximated by the HDMR component functions with unknown coefficients. Then we assign the GP prior for the HDMR model and its partial derivative function. The auto-covariance functions and the cross-covariance functions of these GP are defined by the HDMR component functions respectively. Given a set of samples (response information and gradient information), the posterior distribution of the model response can be computed by means of Bayes theorem, and the GE-HDMR surrogate model is analytically derived by its mean function. The proposed method combines the sample information and gradient information to approximate the response function, thus it provides much more accurate prediction result compared to classic HDMR model with no gradient information. Several examples are used to validate the accuracy and efficiency of the presented method, results demonstrate that the developed GE-HDMR surrogate model is much more efficient and accurate than the classic HDMR model.

The rest of this paper is organized as follows: Section 2 reviews the basic theories of classic HDMR. The construction of the GE-HDMR surrogate model is introduced in detail in Section 3. In Section 4, several tested examples are used to illustrate the performance of the proposed GE-HDMR method. Finally, some conclusions are drawn in Section 5.

Section snippets

High dimensional model representation

High-dimensional function estimation faces with the so-called “curse of dimensionality” as the sample size needed to approximate the function to a satisfying accuracy level increases exponentially with the dimensionality of the function [34]. However, HDMR provides a remarkable way to overcome this predicament by approximating high-dimensional function with a sum of low-dimensional functions [27], [30]. Considering a square-integrable response function y=g(x), n-dimensional independent input

Bayesian inference method

In the Bayesian framework, the crux is to select an appropriate prior for the model output g(x). In general, the GP is always chosen as the prior for g(x) [2]. Given the training samples set D=(X,Y), where X=x1,,xNT is the input data, Y=g(x1),,g(xN)T is the corresponding model response, N is the size of the training samples set, the joint distribution of Y is Gaussian under the GP prior hypothesis, i.e., YN(μY,K), and its joint probability density function (PDF) f(Y) can be defined as f(Y)=(2

Numerical examples

This section is dedicated to the validation and assessment of the proposed GE-HDMR model. In this paper, the HDMR basis functions is chosen to be Hermite polynomials, since they possess the convenient property that they and their derivatives are orthogonal with respect to the same measure [40]. To validate the performance of the GE-HDMR method, various nonlinear mathematical functions and engineering problems are employed, and the corresponding descriptions are listed in Table 1.

To demonstrate

Conclusions

In this paper, we investigated high dimensional model representation surrogate model when gradient information of a response function is present. Assuming the response function and its partial derivative functions are all Gaussian Processes, we developed the auto-covariance functions and cross-covariance functions of all these random process. Then the analytical expression of the GE-HDMR surrogate model is derived from the joint distribution of samples and gradients. The proposed method

Declaration of Competing Interest

No author associated with this paper has disclosed any potential or pertinent conflicts which may be perceived to have impending conflict with this work. For full disclosure statements refer to https://doi.org/10.1016/j.knosys.2019.104903.

Acknowledgments

The authors would like to express the gratitude to three reviewers for helpful comments and constructive suggestions. This work was supported by the National Natural Science Foundation of China (Grant No. NSFC 51775439), National Science and Technology Major Project (2017-IV-0009-0046) and “Innovation Foundation for Doctor Dissertation of Northwestern Polytechnical University” with project code of CX201933.

References (43)

Cited by (0)

View full text