Asymptotic Risk of Least Squares Minimum Norm Estimator under the Spike Covariance Model

12/31/2019
by   Yasaman Mahdaviyeh, et al.
4

One of the recent approaches to explain good performance of neural networks has focused on their ability to fit training data perfectly (interpolate) without overfitting. It has been shown that this is not unique to neural nets, and that it happens with simpler models such as kernel regression, too Belkin et al. (2018b); Tengyuan Liang (2018). Consequently, there has been quite a few works that give conditions for low risk or optimality of interpolating models, see for example Belkin et al. (2018a, 2019b). One of the simpler models where interpolation has been studied recently is least squares solution for linear regression. In this case, interpolation is only guaranteed to happen in high dimensional setting where the number of parameters exceeds number of samples; therefore, least squares solution is not necessarily unique. However, minimum norm solution is unique, can be written in closed form, and gradient descent starting at the origin converges to it (Hastie et al., 2019). This has, at least partially, motivated several works that study risk of minimum norm least squares estimator for linear regression. Continuing in a similar vein, we study the asymptotic risk of minimum norm least squares estimator when number of parameters d depends on n, and d/n→∞. In this high dimensional setting, to make inference feasible, it is usually assumed that true parameters or data have some underlying low dimensional structure such as sparsity, or vanishing eigenvalues of population covariance matrix. Here, we restrict ourselves to spike covariance matrices, where a fixed finite number of eigenvalues grow with n and are much larger than the rest of the eigenvalues, which are (asymptotically) in the same order. We show that in this setting the risk can vanish.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/08/2021

Optimistic Rates: A Unifying Theory for Interpolation Learning and Regularization in Linear Regression

We study a localized notion of uniform convergence known as an "optimist...
research
06/28/2017

Asymptotic Confidence Regions for High-dimensional Structured Sparsity

In the setting of high-dimensional linear regression models, we propose ...
research
04/08/2023

Benign Overfitting of Non-Sparse High-Dimensional Linear Regression with Correlated Noise

We investigate the high-dimensional linear regression problem in situati...
research
02/20/2018

Detection limits in the high-dimensional spiked rectangular model

We study the problem of detecting the presence of a single unknown spike...
research
10/21/2021

On Optimal Interpolation In Linear Regression

Understanding when and why interpolating methods generalize well has rec...
research
02/06/2020

Interpolation under latent factor regression models

This work studies finite-sample properties of the risk of the minimum-no...
research
11/23/2020

Dimensionality reduction, regularization, and generalization in overparameterized regressions

Overparameterization in deep learning is powerful: Very large models fit...

Please sign up or login with your details

Forgot password? Click here to reset