Over-parameterization: A Necessary Condition for Models that Extrapolate

03/20/2022
by   Roozbeh Yousefzadeh, et al.
3

In this work, we study over-parameterization as a necessary condition for having the ability for the models to extrapolate outside the convex hull of training set. We specifically, consider classification models, e.g., image classification and other applications of deep learning. Such models are classification functions that partition their domain and assign a class to each partition <cit.>. Partitions are defined by decision boundaries and so is the classification model/function. Convex hull of training set may occupy only a subset of the domain, but trained model may partition the entire domain and not just the convex hull of training set. This is important because many of the testing samples may be outside the convex hull of training set and the way in which a model partitions its domain outside the convex hull would be influential in its generalization. Using approximation theory, we prove that over-parameterization is a necessary condition for having control over the partitioning of the domain outside the convex hull of training set. We also propose a more clear definition for the notion of over-parametrization based on the learning task and the training set at hand. We present empirical evidence about geometry of datasets, both image and non-image, to provide insights about the extent of extrapolation performed by the models. We consider a 64-dimensional feature space learned by a ResNet model and investigate the geometric arrangements of convex hulls and decision boundaries in that space. We also formalize the notion of extrapolation and relate it to the scope of the model. Finally, we review the rich extrapolation literature in pure and applied mathematics, e.g., the Whitney's Extension Problem, and place our theory in that context.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/25/2021

Deep Learning Generalization and the Convex Hull of Training Sets

We study the generalization of deep learning models in relation to the c...
research
12/06/2021

Extrapolation Frameworks in Cognitive Psychology Suitable for Study of Image Classification Models

We study the functional task of deep learning image classification model...
research
02/05/2022

Decision boundaries and convex hulls in the feature space that deep learning functions learn from images

The success of deep neural networks in image classification and learning...
research
10/01/2019

An Adaptive Sampling Approach for the Reduced Basis Method

The offline time of the reduced basis method can be very long given a la...
research
02/24/2020

Using wavelets to analyze similarities in image datasets

Deep learning image classifiers usually rely on huge training sets and t...
research
07/18/2022

Interpolation, extrapolation, and local generalization in common neural networks

There has been a long history of works showing that neural networks have...
research
06/16/2018

Binary Classification in Unstructured Space With Hypergraph Case-Based Reasoning

Binary classification is one of the most common problem in machine learn...

Please sign up or login with your details

Forgot password? Click here to reset