Conjugate Gradients for Kernel Machines

11/14/2019
by   Simon Bartels, et al.
41

Regularized least-squares (kernel-ridge / Gaussian process) regression is a fundamental algorithm of statistics and machine learning. Because generic algorithms for the exact solution have cubic complexity in the number of datapoints, large datasets require to resort to approximations. In this work, the computation of the least-squares prediction is itself treated as a probabilistic inference problem. We propose a structured Gaussian regression model on the kernel function that uses projections of the kernel matrix to obtain a low-rank approximation of the kernel and the matrix. A central result is an enhanced way to use the method of conjugate gradients for the specific setting of least-squares regression as encountered in machine learning. Our method improves the approximation of the kernel ridge regressor / Gaussian process posterior mean over vanilla conjugate gradients and, allows computation of the posterior variance and the log marginal likelihood (evidence) without further overhead.

READ FULL TEXT

page 4

page 8

research
01/04/2021

Gauss-Legendre Features for Gaussian Process Regression

Gaussian processes provide a powerful probabilistic kernel learning fram...
research
09/11/2016

On the Relationship between Online Gaussian Process Regression and Kernel Least Mean Squares Algorithms

We study the relationship between online Gaussian process (GP) regressio...
research
06/11/2022

Gradient Boosting Performs Low-Rank Gaussian Process Inference

This paper shows that gradient boosting based on symmetric decision tree...
research
12/17/2013

The Matrix Ridge Approximation: Algorithms and Applications

We are concerned with an approximation problem for a symmetric positive ...
research
02/22/2016

Preconditioning Kernel Matrices

The computational and storage complexity of kernel machines presents the...
research
02/22/2022

Adaptive Cholesky Gaussian Processes

We present a method to fit exact Gaussian process models to large datase...
research
10/29/2011

Efficient Marginal Likelihood Computation for Gaussian Process Regression

In a Bayesian learning setting, the posterior distribution of a predicti...

Please sign up or login with your details

Forgot password? Click here to reset