Cross-validation based adaptive sampling for Gaussian process models

05/04/2020
by   Hossein Mohammadi, et al.
0

In many real-world applications, we are interested in approximating black-box, costly functions as accurately as possible with the smallest number of function evaluations. A complex computer code is an example of such a function. In this work, a Gaussian process (GP) emulator is used to approximate the output of complex computer code. We consider the problem of extending an initial experiment sequentially to improve the emulator. A sequential sampling approach based on leave-one-out (LOO) cross-validation is proposed that can be easily extended to a batch mode. This is a desirable property since it saves the user time when parallel computing is available. After fitting a GP to training data points, the expected squared LOO error (ESE_LOO) is calculated at each design point. ESE_LOO is used as a measure to identify important data points. More precisely, when this quantity is large at a point it means that the quality of prediction depends a great deal on that point and adding more samples in the nearby region could improve the accuracy of the GP model. As a result, it is reasonable to select the next sample where ESE_LOO is maximum. However, such quantity is only known at the experimental design and needs to be estimated at unobserved points. To do this, a second GP is fitted to the ESE_LOOs and where the maximum of the modified expected improvement (EI) criterion occurs is chosen as the next sample. EI is a popular acquisition function in Bayesian optimisation and is used to trade-off between local/global search. However, it has tendency towards exploitation, meaning that its maximum is close to the (current) "best" sample. To avoid clustering, a modified version of EI, called pseudo expected improvement, is employed which is more explorative than EI and allows us to discover unexplored regions. The results show that the proposed sampling method is promising.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/18/2023

Cross-Validation Based Adaptive Sampling for Multi-Level Gaussian Process Models

Complex computer codes or models can often be run in a hierarchy of diff...
research
10/12/2019

Bayesian Optimization using Pseudo-Points

Bayesian optimization (BO) is a popular approach for expensive black-box...
research
06/24/2022

Sequential adaptive design for emulating costly computer codes

Gaussian processes (GPs) are generally regarded as the gold standard sur...
research
10/21/2020

Batch Sequential Adaptive Designs for Global Optimization

Compared with the fixed-run designs, the sequential adaptive designs (SA...
research
03/08/2016

Small ensembles of kriging models for optimization

The Efficient Global Optimization (EGO) algorithm uses a conditional Gau...
research
07/26/2018

Deriving Information Acquisition Criteria For Sequentially Inferring The Expected Value Of A Black-Box Function

Acquiring information about noisy expensive black-box functions (compute...
research
05/30/2023

Parallelized Acquisition for Active Learning using Monte Carlo Sampling

Bayesian inference remains one of the most important tool-kits for any s...

Please sign up or login with your details

Forgot password? Click here to reset