An informational approach to the global optimization of expensive-to-evaluate functions
Julien Villemonteix, Emmanuel Vazquez, Eric Walter
TL;DR
The paper tackles global optimization with expensive-to-evaluate functions by combining Kriging surrogates with an information-theoretic criterion. It introduces minimizer entropy and the IAGO algorithm within a SUR framework to sequentially choose evaluation points, demonstrating substantial evaluation savings over EI-based methods and extending to robust, noisy settings. Through conditional simulations and a probabilistic view of the global minimizers, IAGO achieves more global exploration and faster uncertainty reduction, with practical relevance for engineering problems where each evaluation is costly. The work also discusses robustness to covariance-parameter uncertainty and outlines avenues for constrained optimization and higher-dimensional applications, highlighting the method's potential for efficient, uncertainty-aware design under noise.
Abstract
In many global optimization problems motivated by engineering applications, the number of function evaluations is severely limited by time or cost. To ensure that each evaluation contributes to the localization of good candidates for the role of global minimizer, a sequential choice of evaluation points is usually carried out. In particular, when Kriging is used to interpolate past evaluations, the uncertainty associated with the lack of information on the function can be expressed and used to compute a number of criteria accounting for the interest of an additional evaluation at any given point. This paper introduces minimizer entropy as a new Kriging-based criterion for the sequential choice of points at which the function should be evaluated. Based on \emph{stepwise uncertainty reduction}, it accounts for the informational gain on the minimizer expected from a new evaluation. The criterion is approximated using conditional simulations of the Gaussian process model behind Kriging, and then inserted into an algorithm similar in spirit to the \emph{Efficient Global Optimization} (EGO) algorithm. An empirical comparison is carried out between our criterion and \emph{expected improvement}, one of the reference criteria in the literature. Experimental results indicate major evaluation savings over EGO. Finally, the method, which we call IAGO (for Informational Approach to Global Optimization) is extended to robust optimization problems, where both the factors to be tuned and the function evaluations are corrupted by noise.
