Log In Sign Up

Posterior Variance Analysis of Gaussian Processes with Application to Average Learning Curves

by   Armin Lederer, et al.

The posterior variance of Gaussian processes is a valuable measure of the learning error which is exploited in various applications such as safe reinforcement learning and control design. However, suitable analysis of the posterior variance which captures its behavior for finite and infinite number of training data is missing. This paper derives a novel bound for the posterior variance function which requires only local information because it depends only on the number of training samples in the proximity of a considered test point. Furthermore, we prove sufficient conditions which ensure the convergence of the posterior variance to zero. Finally, we demonstrate that the extension of our bound to an average learning bound outperforms existing approaches.


page 1

page 2

page 3

page 4


Uniform Error and Posterior Variance Bounds for Gaussian Process Regression with Application to Safe Control

In application areas where data generation is expensive, Gaussian proces...

Posterior and Computational Uncertainty in Gaussian Processes

Gaussian processes scale prohibitively with the size of the dataset. In ...

Upgrading from Gaussian Processes to Student's-T Processes

Gaussian process priors are commonly used in aerospace design for perfor...

Bayesian Active Learning with Fully Bayesian Gaussian Processes

The bias-variance trade-off is a well-known problem in machine learning ...

Uniform Error Bounds for Gaussian Process Regression with Application to Safe Control

Data-driven models are subject to model errors due to limited and noisy ...

Gaussian Processes Over Graphs

We propose Gaussian processes for signals over graphs (GPG) using the ap...

Asymmetric kernel in Gaussian Processes for learning target variance

This work incorporates the multi-modality of the data distribution into ...