A finite sample analysis of the double descent phenomenon for ridge function estimation

07/25/2020
by   Emmanuel Caron, et al.
0

Recent extensive numerical experiments in high scale machine learning have allowed to uncover a quite counterintuitive phase transition, as a function of the ratio between the sample size and the number of parameters in the model. As the number of parameters p approaches the sample size n, the generalisation error (a.k.a. testing error) increases, but it many cases, it starts decreasing again past the threshold p=n. This surprising phenomenon, brought to the theoretical community attention in <cit.>, has been thorougly investigated lately, more specifically for simpler models than deep neural networks, such as the linear model when the parameter is taken to be the minimum norm solution to the least-square problem, mostly in the asymptotic regime when p and n tend to +∞; see e.g. <cit.>. In the present paper, we propose a finite sample analysis of non-linear models of ridge type, where we investigate the double descent phenomenon for both the estimation problem and the prediction problem. Our results show that the double descent phenomenon can be precisely demonstrated in non-linear settings and complements recent works of <cit.> and <cit.>. Our analysis is based on efficient but elementary tools closely related to the continuous Newton method <cit.>.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/08/2021

Asymptotics of Ridge Regression in Convolutional Models

Understanding generalization and estimation error of estimators for simp...
research
08/03/2023

Functional Data Regression Reconciles with Excess Bases

As the development of measuring instruments and computers has accelerate...
research
12/10/2019

Exact expressions for double descent and implicit regularization via surrogate random design

Double descent refers to the phase transition that is exhibited by the g...
research
04/17/2023

Analysis of Interpolating Regression Models and the Double Descent Phenomenon

A regression model with more parameters than data points in the training...
research
10/18/2021

Minimum ℓ_1-norm interpolators: Precise asymptotics and multiple descent

An evolving line of machine learning works observe empirical evidence th...
research
05/13/2022

Sharp Asymptotics of Kernel Ridge Regression Beyond the Linear Regime

The generalization performance of kernel ridge regression (KRR) exhibits...
research
03/14/2023

Testing Causality for High Dimensional Data

Determining causal relationship between high dimensional observations ar...

Please sign up or login with your details

Forgot password? Click here to reset