Dimensionality reduction, regularization, and generalization in overparameterized regressions

11/23/2020
by   Ningyuan Huang, et al.
0

Overparameterization in deep learning is powerful: Very large models fit the training data perfectly and yet generalize well. This realization brought back the study of linear models for regression, including ordinary least squares (OLS), which, like deep learning, shows a "double descent" behavior. This involves two features: (1) The risk (out-of-sample prediction error) can grow arbitrarily when the number of samples n approaches the number of parameters p, and (2) the risk decreases with p at p>n, sometimes achieving a lower value than the lowest risk at p<n. The divergence of the risk for OLS at p≈ n is related to the condition number of the empirical covariance in the feature set. For this reason, it can be avoided with regularization. In this work we show that it can also be avoided with a PCA-based dimensionality reduction. We provide a finite upper bound for the risk of the PCA-based estimator. This result is in contrast to recent work that shows that a different form of dimensionality reduction – one based on the population covariance instead of the empirical covariance – does not avoid the divergence. We connect these results to an analysis of adversarial attacks, which become more effective as they raise the condition number of the empirical covariance of the features. We show that OLS is arbitrarily susceptible to data-poisoning attacks in the overparameterized regime – unlike the underparameterized regime – and that regularization and dimensionality reduction improve the robustness.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/18/2020

The Dilemma Between Dimensionality Reduction and Adversarial Robustness

Recent work has shown the tremendous vulnerability to adversarial sample...
research
09/29/2015

Foundations of Coupled Nonlinear Dimensionality Reduction

In this paper we introduce and analyze the learning scenario of coupled ...
research
12/19/2022

Prediction of Auto Insurance Risk Based on t-SNE Dimensionality Reduction

Correct scoring of a driver's risk is of great significance to auto insu...
research
04/08/2022

Dimensionality Reduction in Deep Learning via Kronecker Multi-layer Architectures

Deep learning using neural networks is an effective technique for genera...
research
02/02/2020

Overfitting Can Be Harmless for Basis Pursuit: Only to a Degree

Recently, there have been significant interests in studying the generali...
research
12/31/2019

Asymptotic Risk of Least Squares Minimum Norm Estimator under the Spike Covariance Model

One of the recent approaches to explain good performance of neural netwo...
research
02/09/2022

Empirical Risk Minimization with Relative Entropy Regularization: Optimality and Sensitivity Analysis

The optimality and sensitivity of the empirical risk minimization proble...

Please sign up or login with your details

Forgot password? Click here to reset