Bias-Free Scalable Gaussian Processes via Randomized Truncations

02/12/2021
by   Andres Potapczynski, et al.
0

Scalable Gaussian Process methods are computationally attractive, yet introduce modeling biases that require rigorous study. This paper analyzes two common techniques: early truncated conjugate gradients (CG) and random Fourier features (RFF). We find that both methods introduce a systematic bias on the learned hyperparameters: CG tends to underfit while RFF tends to overfit. We address these issues using randomized truncation estimators that eliminate bias in exchange for increased variance. In the case of RFF, we show that the bias-to-variance conversion is indeed a trade-off: the additional variance proves detrimental to optimization. However, in the case of CG, our unbiased learning procedure meaningfully outperforms its biased counterpart with minimal additional computation.

READ FULL TEXT
research
09/20/2021

Barely Biased Learning for Gaussian Process Regression

Recent work in scalable approximate Gaussian process regression has disc...
research
05/23/2022

Bayesian Active Learning with Fully Bayesian Gaussian Processes

The bias-variance trade-off is a well-known problem in machine learning ...
research
12/27/2021

Unbiased Gradient Estimation in Unrolled Computation Graphs with Persistent Evolution Strategies

Unrolled computation graphs arise in many scenarios, including training ...
research
09/03/2023

Double Clipping: Less-Biased Variance Reduction in Off-Policy Evaluation

"Clipping" (a.k.a. importance weight truncation) is a widely used varian...
research
01/22/2015

Enabling scalable stochastic gradient-based inference for Gaussian processes by employing the Unbiased LInear System SolvEr (ULISSE)

In applications of Gaussian processes where quantification of uncertaint...
research
07/01/2021

Reducing the Variance of Gaussian Process Hyperparameter Optimization with Preconditioning

Gaussian processes remain popular as a flexible and expressive model cla...
research
06/09/2022

On the Bias-Variance Characteristics of LIME and SHAP in High Sparsity Movie Recommendation Explanation Tasks

We evaluate two popular local explainability techniques, LIME and SHAP, ...

Please sign up or login with your details

Forgot password? Click here to reset