Bandwidth Selection for Gaussian Kernel Ridge Regression via Jacobian Control

05/24/2022
by   Oskar Allerbo, et al.
0

Most machine learning methods depend on the tuning of hyper-parameters. For kernel ridge regression (KRR) with the Gaussian kernel, the hyper-parameter is the bandwidth. The bandwidth specifies the length-scale of the kernel and has to be carefully selected in order to obtain a model with good generalization. The default method for bandwidth selection is cross-validation, which often yields good results, albeit at high computational costs. Furthermore, the estimates provided by cross-validation tend to have very high variance, especially when training data are scarce. Inspired by Jacobian regularization, we formulate how the derivatives of the functions inferred by KRR with the Gaussian kernel depend on the kernel bandwidth. We then use this expression to propose a closed-form, computationally feather-light, bandwidth selection method based on controlling the Jacobian. In addition, the Jacobian expression illuminates how the bandwidth selection is a trade-off between the smoothness of the inferred function, and the conditioning of the training data kernel matrix. We show on real and synthetic data that compared to cross-validation, our method is considerably more stable in terms of bandwidth selection, and, for small data sets, provides better predictions.

READ FULL TEXT
research
01/26/2020

On a Nadaraya-Watson Estimator with Two Bandwidths

In a regression model, we write the Nadaraya-Watson estimator of the reg...
research
06/29/2023

Solving Kernel Ridge Regression with Gradient-Based Optimization Methods

Kernel ridge regression, KRR, is a non-linear generalization of linear r...
research
06/17/2020

Revisiting complexity and the bias-variance tradeoff

The recent success of high-dimensional models, such as deep neural netwo...
research
01/19/2023

Data-driven kernel designs for optimized greedy schemes: A machine learning perspective

Thanks to their easy implementation via Radial Basis Functions (RBFs), m...
research
04/16/2022

FKreg: A MATLAB toolbox for fast Multivariate Kernel Regression

Kernel smooth is the most fundamental technique for data density and reg...
research
09/10/2009

Data-driven calibration of linear estimators with minimal penalties

This paper tackles the problem of selecting among several linear estimat...
research
10/07/2020

A consistent second-order discrete kernel smoother

The histogram estimator of a discrete probability mass function often ex...

Please sign up or login with your details

Forgot password? Click here to reset