Interpolation can hurt robust generalization even when there is no noise

08/05/2021
by   Konstantin Donhauser, et al.
0

Numerous recent works show that overparameterization implicitly reduces variance for min-norm interpolators and max-margin classifiers. These findings suggest that ridge regularization has vanishing benefits in high dimensions. We challenge this narrative by showing that, even in the absence of noise, avoiding interpolation through ridge regularization can significantly improve generalization. We prove this phenomenon for the robust risk of both linear regression and classification and hence provide the first theoretical result on robust overfitting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/17/2022

Beyond Ridge Regression for Distribution-Free Data

In supervised batch learning, the predictive normalized maximum likeliho...
research
03/23/2021

Benign Overfitting of Constant-Stepsize SGD for Linear Regression

There is an increasing realization that algorithmic inductive biases are...
research
07/05/2023

The distribution of Ridgeless least squares interpolators

The Ridgeless minimum ℓ_2-norm interpolator in overparametrized linear r...
research
09/29/2020

Benign overfitting in ridge regression

Classical learning theory suggests that strong regularization is needed ...
research
06/05/2020

Triple descent and the two kinds of overfitting: Where why do they appear?

A recent line of research has highlighted the existence of a double desc...
research
11/28/2022

Malign Overfitting: Interpolation Can Provably Preclude Invariance

Learned classifiers should often possess certain invariance properties m...
research
05/07/2020

Fractional ridge regression: a fast, interpretable reparameterization of ridge regression

Ridge regression (RR) is a regularization technique that penalizes the L...

Please sign up or login with your details

Forgot password? Click here to reset