Testing for Overfitting

05/09/2023
by   James Schmidt, et al.
0

High complexity models are notorious in machine learning for overfitting, a phenomenon in which models well represent data but fail to generalize an underlying data generating process. A typical procedure for circumventing overfitting computes empirical risk on a holdout set and halts once (or flags that/when) it begins to increase. Such practice often helps in outputting a well-generalizing model, but justification for why it works is primarily heuristic. We discuss the overfitting problem and explain why standard asymptotic and concentration results do not hold for evaluation with training data. We then proceed to introduce and argue for a hypothesis test by means of which both model performance may be evaluated using training data, and overfitting quantitatively defined and detected. We rely on said concentration bounds which guarantee that empirical means should, with high probability, approximate their true mean to conclude that they should approximate each other. We stipulate conditions under which this test is valid, describe how the test may be used for identifying overfitting, articulate a further nuance according to which distributional shift may be flagged, and highlight an alternative notion of learning which usefully captures generalization in the absence of uniform PAC guarantees.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/01/2020

The Restricted Isometry of ReLU Networks: Generalization through Norm Concentration

While regression tasks aim at interpolating a relation on the entire inp...
research
07/01/2016

A new analytical approach to consistency and overfitting in regularized empirical risk minimization

This work considers the problem of binary classification: given training...
research
01/04/2021

Benign overfitting without concentration

We obtain a sufficient condition for benign overfitting of linear regres...
research
06/11/2020

A new measure for overfitting and its implications for backdooring of deep learning

Overfitting describes the phenomenon that a machine learning model fits ...
research
05/30/2023

Quantifying Overfitting: Evaluating Neural Network Performance through Analysis of Null Space

Machine learning models that are overfitted/overtrained are more vulnera...
research
08/25/2021

The Interplay Between Implicit Bias and Benign Overfitting in Two-Layer Linear Networks

The recent success of neural network models has shone light on a rather ...
research
10/12/2015

Toward a Better Understanding of Leaderboard

The leaderboard in machine learning competitions is a tool to show the p...

Please sign up or login with your details

Forgot password? Click here to reset