Small ensembles of kriging models for optimization

03/08/2016
by   Hossein Mohammadi, et al.
0

The Efficient Global Optimization (EGO) algorithm uses a conditional Gaus-sian Process (GP) to approximate an objective function known at a finite number of observation points and sequentially adds new points which maximize the Expected Improvement criterion according to the GP. The important factor that controls the efficiency of EGO is the GP covariance function (or kernel) which should be chosen according to the objective function. Traditionally, a pa-rameterized family of covariance functions is considered whose parameters are learned through statistical procedures such as maximum likelihood or cross-validation. However, it may be questioned whether statistical procedures for learning covariance functions are the most efficient for optimization as they target a global agreement between the GP and the observations which is not the ultimate goal of optimization. Furthermore, statistical learning procedures are computationally expensive. The main alternative to the statistical learning of the GP is self-adaptation, where the algorithm tunes the kernel parameters based on their contribution to objective function improvement. After questioning the possibility of self-adaptation for kriging based optimizers, this paper proposes a novel approach for tuning the length-scale of the GP in EGO: At each iteration, a small ensemble of kriging models structured by their length-scales is created. All of the models contribute to an iterate in an EGO-like fashion. Then, the set of models is densified around the model whose length-scale yielded the best iterate and further points are produced. Numerical experiments are provided which motivate the use of many length-scales. The tested implementation does not perform better than the classical EGO algorithm in a sequential context but show the potential of the approach for parallel implementations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/14/2023

Comparing Scale Parameter Estimators for Gaussian Process Regression: Cross Validation and Maximum Likelihood

Gaussian process (GP) regression is a Bayesian nonparametric method for ...
research
04/29/2021

MuyGPs: Scalable Gaussian Process Hyperparameter Estimation Using Local Cross-Validation

Gaussian processes (GPs) are non-linear probabilistic models popular in ...
research
05/04/2020

Cross-validation based adaptive sampling for Gaussian process models

In many real-world applications, we are interested in approximating blac...
research
06/07/2022

Relaxed Gaussian process interpolation: a goal-oriented approach to Bayesian optimization

This work presents a new procedure for obtaining predictive distribution...
research
07/07/2020

srMO-BO-3GP: A sequential regularized multi-objective constrained Bayesian optimization for design applications

Bayesian optimization (BO) is an efficient and flexible global optimizat...
research
07/18/2023

Cross-Validation Based Adaptive Sampling for Multi-Level Gaussian Process Models

Complex computer codes or models can often be run in a hierarchy of diff...
research
07/20/2020

Iterative Method for Tuning Complex Simulation Code

Tuning a complex simulation code refers to the process of improving the ...

Please sign up or login with your details

Forgot password? Click here to reset