Fast calculation of Gaussian Process multiple-fold cross-validation residuals and their covariances

01/08/2021
by   David Ginsbourger, et al.
0

We generalize fast Gaussian process leave-one-out formulae to multiple-fold cross-validation, highlighting in turn in broad settings the covariance structure of cross-validation residuals. The employed approach, that relies on block matrix inversion via Schur complements, is applied to both Simple and Universal Kriging frameworks. We illustrate how resulting covariances affect model diagnostics and how to properly transform residuals in the first place. Beyond that, we examine how accounting for dependency between such residuals affect cross-validation-based estimation of the scale parameter. It is found in two distinct cases, namely in scale estimation and in broader covariance parameter estimation via pseudo-likelihood, that correcting for covariances between cross-validation residuals leads back to maximum likelihood estimation or to an original variation thereof. The proposed fast calculation of Gaussian Process multiple-fold cross-validation residuals is implemented and benchmarked against a naive implementation, all in R language. Numerical experiments highlight the accuracy of our approach as well as the substantial speed-ups that it enables. It is noticeable however, as supported by a discussion on the main drivers of computational costs and by a dedicated numerical benchmark, that speed-ups steeply decline as the number of folds (say, all sharing the same size) decreases. Overall, our results enable fast multiple-fold cross-validation, have direct consequences in GP model diagnostics, and pave the way to future work on hyperparameter fitting as well as on the promising field of goal-oriented fold design.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/14/2023

Comparing Scale Parameter Estimators for Gaussian Process Regression: Cross Validation and Maximum Likelihood

Gaussian process (GP) regression is a Bayesian nonparametric method for ...
research
02/26/2020

Towards new cross-validation-based estimators for Gaussian process regression: efficient adjoint computation of gradients

We consider the problem of estimating the parameters of the covariance f...
research
03/21/2021

Detecting Label Noise via Leave-One-Out Cross-Validation

We present a simple algorithm for identifying and correcting real-valued...
research
08/22/2019

Efficient Cross-Validation of Echo State Networks

Echo State Networks (ESNs) are known for their fast and precise one-shot...
research
11/12/2018

An Easy Implementation of CV-TMLE

In the world of targeted learning, cross-validated targeted maximum like...
research
11/21/2021

A stochastic extended Rippa's algorithm for LpOCV

In kernel-based approximation, the tuning of the so-called shape paramet...
research
12/31/2019

Parallel cross-validation: a scalable fitting method for Gaussian process models

Gaussian process (GP) models are widely used to analyze spatially refere...

Please sign up or login with your details

Forgot password? Click here to reset