Risk Convergence of Centered Kernel Ridge Regression with Large Dimensional Data

04/19/2019
by   Khalil Elkhalil, et al.
0

This paper carries out a large dimensional analysis of a variation of kernel ridge regression that we call centered kernel ridge regression (CKRR), also known in the literature as kernel ridge regression with offset. This modified technique is obtained by accounting for the bias in the regression problem resulting in the old kernel ridge regression but with centered kernels. The analysis is carried out under the assumption that the data is drawn from a Gaussian distribution and heavily relies on tools from random matrix theory (RMT). Under the regime in which the data dimension and the training size grow infinitely large with fixed ratio and under some mild assumptions controlling the data statistics, we show that both the empirical and the prediction risks converge to a deterministic quantities that describe in closed form fashion the performance of CKRR in terms of the data statistics and dimensions. Inspired by this theoretical result, we subsequently build a consistent estimator of the prediction risk based on the training data which allows to optimally tune the design parameters. A key insight of the proposed analysis is the fact that asymptotically a large class of kernels achieve the same minimum prediction risk. This insight is validated with both synthetic and real data.

READ FULL TEXT
research
06/17/2020

Kernel Alignment Risk Estimator: Risk Prediction from Training Data

We study the risk (i.e. generalization error) of Kernel Ridge Regression...
research
02/07/2022

Failure and success of the spectral bias prediction for Kernel Ridge Regression: the case of low-dimensional data

Recently, several theories including the replica method made predictions...
research
04/09/2021

How rotational invariance of common kernels prevents generalization in high dimensions

Kernel ridge regression is well-known to achieve minimax optimal rates i...
research
09/17/2018

Statistically and Computationally Efficient Variance Estimator for Kernel Ridge Regression

In this paper, we propose a random projection approach to estimate varia...
research
05/29/2023

Generalized equivalences between subsampling and ridge regularization

We establish precise structural and risk equivalences between subsamplin...
research
06/09/2018

Deterministic Stretchy Regression

An extension of the regularized least-squares in which the estimation pa...
research
09/10/2023

Nonlinear Granger Causality using Kernel Ridge Regression

I introduce a novel algorithm and accompanying Python library, named mlc...

Please sign up or login with your details

Forgot password? Click here to reset