Interpolation, extrapolation, and local generalization in common neural networks

07/18/2022
by   Laurent Bonnasse-Gahot, et al.
0

There has been a long history of works showing that neural networks have hard time extrapolating beyond the training set. A recent study by Balestriero et al. (2021) challenges this view: defining interpolation as the state of belonging to the convex hull of the training set, they show that the test set, either in input or neural space, cannot lie for the most part in this convex hull, due to the high dimensionality of the data, invoking the well known curse of dimensionality. Neural networks are then assumed to necessarily work in extrapolative mode. We here study the neural activities of the last hidden layer of typical neural networks. Using an autoencoder to uncover the intrinsic space underlying the neural activities, we show that this space is actually low-dimensional, and that the better the model, the lower the dimensionality of this intrinsic space. In this space, most samples of the test set actually lie in the convex hull of the training set: under the convex hull definition, the models thus happen to work in interpolation regime. Moreover, we show that belonging to the convex hull does not seem to be the relevant criteria. Different measures of proximity to the training set are actually better related to performance accuracy. Thus, typical neural networks do seem to operate in interpolation regime. Good generalization performances are linked to the ability of a neural network to operate well in such a regime.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/25/2020

The Interpolation Phase Transition in Neural Networks: Memorization and Generalization under Lazy Training

Modern neural networks are often operated in a strongly overparametrized...
research
01/25/2021

Deep Learning Generalization and the Convex Hull of Training Sets

We study the generalization of deep learning models in relation to the c...
research
09/25/2019

Regularising Deep Networks with DGMs

Here we develop a new method for regularising neural networks where we l...
research
06/28/2020

Modeling Generalization in Machine Learning: A Methodological and Computational Study

As machine learning becomes more and more available to the general publi...
research
12/18/2022

The Underlying Correlated Dynamics in Neural Training

Training of neural networks is a computationally intensive task. The sig...
research
03/30/2018

Learning to generate classifiers

We train a network to generate mappings between training sets and classi...
research
03/20/2022

Over-parameterization: A Necessary Condition for Models that Extrapolate

In this work, we study over-parameterization as a necessary condition fo...

Please sign up or login with your details

Forgot password? Click here to reset