DeepAI AI Chat
Log In Sign Up

Approximate Inference for Nonstationary Heteroscedastic Gaussian process Regression

by   Ville Tolvanen, et al.

This paper presents a novel approach for approximate integration over the uncertainty of noise and signal variances in Gaussian process (GP) regression. Our efficient and straightforward approach can also be applied to integration over input dependent noise variance (heteroscedasticity) and input dependent signal variance (nonstationarity) by setting independent GP priors for the noise and signal variances. We use expectation propagation (EP) for inference and compare results to Markov chain Monte Carlo in two simulated data sets and three empirical examples. The results show that EP produces comparable results with less computational burden.


page 1

page 2

page 3

page 4


Gaussian Process Regression with Heteroscedastic or Non-Gaussian Residuals

Gaussian Process (GP) regression models typically assume that residuals ...

Nested Expectation Propagation for Gaussian Process Classification with a Multinomial Probit Likelihood

We consider probabilistic multinomial probit classification using Gaussi...

Gaussian Process Regression Networks

We introduce a new regression framework, Gaussian process regression net...

Sparse Additive Gaussian Process Regression

In this paper we introduce a novel model for Gaussian process (GP) regre...

Non-Stationary Gaussian Process Regression with Hamiltonian Monte Carlo

We present a novel approach for fully non-stationary Gaussian process re...

Expectation Propagation for Neural Networks with Sparsity-promoting Priors

We propose a novel approach for nonlinear regression using a two-layer n...

Variance Reduction for Metropolis-Hastings Samplers

We introduce a general framework that constructs estimators with reduced...