Discretization-free Knowledge Gradient Methods for Bayesian Optimization

07/20/2017
by   Jian Wu, et al.
0

This paper studies Bayesian ranking and selection (R&S) problems with correlated prior beliefs and continuous domains, i.e. Bayesian optimization (BO). Knowledge gradient methods [Frazier et al., 2008, 2009] have been widely studied for discrete R&S problems, which sample the one-step Bayes-optimal point. When used over continuous domains, previous work on the knowledge gradient [Scott et al., 2011, Wu and Frazier, 2016, Wu et al., 2017] often rely on a discretized finite approximation. However, the discretization introduces error and scales poorly as the dimension of domain grows. In this paper, we develop a fast discretization-free knowledge gradient method for Bayesian optimization. Our method is not restricted to the fully sequential setting, but useful in all settings where knowledge gradient can be used over continuous domains. We show how our method can be generalized to handle (i) batch of points suggestion (parallel knowledge gradient); (ii) the setting where derivative information is available in the optimization process (derivative-enabled knowledge gradient). In numerical experiments, we demonstrate that the discretization-free knowledge gradient method finds global optima significantly faster than previous Bayesian optimization algorithms on both synthetic test functions and real-world applications, especially when function evaluations are noisy; and derivative-enabled knowledge gradient can further improve the performances, even outperforming the gradient-based optimizer such as BFGS when derivative information is available.

READ FULL TEXT
research
06/14/2016

The Parallel Knowledge Gradient Method for Batch Bayesian Optimization

In many applications of black-box optimization, one can evaluate multipl...
research
03/13/2017

Bayesian Optimization with Gradients

Bayesian optimization has been successful at global optimization of expe...
research
05/29/2019

Linear interpolation gives better gradients than Gaussian smoothing in derivative-free optimization

In this paper, we consider derivative free optimization problems, where ...
research
02/16/2016

Parallel Bayesian Global Optimization of Expensive Functions

We consider parallel global optimization of derivative-free expensive-to...
research
08/01/2023

Hessian-Aware Bayesian Optimization for Decision Making Systems

Many approaches for optimizing decision making systems rely on gradient ...
research
03/04/2022

Distributionally Robust Bayesian Optimization with φ-divergences

The study of robustness has received much attention due to its inevitabi...
research
07/05/2021

DIRECTGO: A new DIRECT-type MATLAB toolbox for derivative-free global optimization

In this work, we introduce DIRECTGO, a new MATLAB toolbox for derivative...

Please sign up or login with your details

Forgot password? Click here to reset