Group-regularized ridge regression via empirical Bayes noise level cross-validation

10/29/2020
by   Nikolaos Ignatiadis, et al.
0

Features in predictive models are not exchangeable, yet common supervised models treat them as such. Here we study ridge regression when the analyst can partition the features into K groups based on external side-information. For example, in high-throughput biology, features may represent gene expression, protein abundance or clinical data and so each feature group represents a distinct modality. The analyst's goal is to choose optimal regularization parameters λ = (λ_1, …, λ_K) – one for each group. In this work, we study the impact of λ on the predictive risk of group-regularized ridge regression by deriving limiting risk formulae under a high-dimensional random effects model with p≍ n as n →∞. Furthermore, we propose a data-driven method for choosing λ that attains the optimal asymptotic risk: The key idea is to interpret the residual noise variance σ^2, as a regularization parameter to be chosen through cross-validation. An empirical Bayes construction maps the one-dimensional parameter σ to the K-dimensional vector of regularization parameters, i.e., σ↦λ(σ). Beyond its theoretical optimality, the proposed method is practical and runs as fast as cross-validated ridge regression without feature groups (K=1).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/06/2019

Ridge Regression: Structure, Cross-Validation, and Sketching

We study the following three fundamental problems about ridge regression...
research
03/31/2017

The Risk of Machine Learning

Many applied settings in empirical economics involve simultaneous estima...
research
05/07/2020

Fractional ridge regression: a fast, interpretable reparameterization of ridge regression

Ridge regression (RR) is a regularization technique that penalizes the L...
research
06/17/2020

Revisiting complexity and the bias-variance tradeoff

The recent success of high-dimensional models, such as deep neural netwo...
research
11/07/2018

Adaptive penalization in high-dimensional regression and classification with external covariates using variational Bayes

Penalization schemes like Lasso or ridge regression are routinely used t...
research
05/19/2020

Fast cross-validation for multi-penalty ridge regression

Prediction based on multiple high-dimensional data types needs to accoun...
research
07/10/2021

Cluster Regularization via a Hierarchical Feature Regression

Prediction tasks with high-dimensional nonorthogonal predictor sets pose...

Please sign up or login with your details

Forgot password? Click here to reset