Marginalised Gaussian Processes with Nested Sampling

10/30/2020
by   Fergus Simpson, et al.
0

Gaussian Process (GPs) models are a rich distribution over functions with inductive biases controlled by a kernel function. Learning occurs through the optimisation of kernel hyperparameters using the marginal likelihood as the objective. This classical approach known as Type-II maximum likelihood (ML-II) yields point estimates of the hyperparameters, and continues to be the default method for training GPs. However, this approach risks underestimating predictive uncertainty and is prone to overfitting especially when there are many hyperparameters. Furthermore, gradient based optimisation makes ML-II point estimates highly susceptible to the presence of local minima. This work presents an alternative learning procedure where the hyperparameters of the kernel function are marginalised using Nested Sampling (NS), a technique that is well suited to sample from complex, multi-modal distributions. We focus on regression tasks with the spectral mixture (SM) class of kernels and find that a principled approach to quantifying model uncertainty leads to substantial gains in predictive performance across a range of synthetic and benchmark data sets. In this context, nested sampling is also found to offer a speed advantage over Hamiltonian Monte Carlo (HMC), widely considered to be the gold-standard in MCMC based inference.

READ FULL TEXT

page 7

page 8

research
12/31/2019

Approximate Inference for Fully Bayesian Gaussian Process Regression

Learning in Gaussian Process models occurs through the adaptation of hyp...
research
05/03/2021

How Bayesian Should Bayesian Optimisation Be?

Bayesian optimisation (BO) uses probabilistic surrogate models - usually...
research
11/04/2022

Sparse Gaussian Process Hyperparameters: Optimize or Integrate?

The kernel function and its hyperparameters are the central model select...
research
10/29/2019

Function-Space Distributions over Kernels

Gaussian processes are flexible function approximators, with inductive b...
research
03/19/2018

Learning non-Gaussian Time Series using the Box-Cox Gaussian Process

Gaussian processes (GPs) are Bayesian nonparametric generative models th...
research
10/11/2022

Computationally-efficient initialisation of GPs: The generalised variogram method

We present a computationally-efficient strategy to find the hyperparamet...
research
02/02/2018

Scalable Lévy Process Priors for Spectral Kernel Learning

Gaussian processes are rich distributions over functions, with generaliz...

Please sign up or login with your details

Forgot password? Click here to reset