Two models of double descent for weak features

03/18/2019
by   Mikhail Belkin, et al.
6

The "double descent" risk curve was recently proposed to qualitatively describe the out-of-sample prediction accuracy of variably-parameterized machine learning models. This article provides a precise mathematical analysis for the shape of this curve in two simple data models with the least squares/least norm predictor. Specifically, it is shown that the risk peaks when the number of features p is close to the sample size n, but also that the risk decreases towards its minimum as p increases beyond n. This behavior is contrasted with that of "prescient" models that select features in an a priori optimal order.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/07/2020

A Brief Prehistory of Double Descent

In their thought-provoking paper [1], Belkin et al. illustrate and discu...
research
08/21/2022

Multiple Descent in the Multiple Random Feature Model

Recent works have demonstrated a double descent phenomenon in over-param...
research
10/18/2021

Minimum ℓ_1-norm interpolators: Precise asymptotics and multiple descent

An evolving line of machine learning works observe empirical evidence th...
research
08/03/2020

Multiple Descent: Design Your Own Generalization Curve

This paper explores the generalization loss of linear regression in vari...
research
11/19/2018

Domain of Inverse Double Arcsine Transformation

To combine the proportions from different studies for meta-analysis, Fre...
research
11/18/2022

Understanding the double descent curve in Machine Learning

The theory of bias-variance used to serve as a guide for model selection...
research
07/27/2021

On the Role of Optimization in Double Descent: A Least Squares Study

Empirically it has been observed that the performance of deep neural net...

Please sign up or login with your details

Forgot password? Click here to reset