Distributional Generalization: A New Kind of Generalization

09/17/2020
by   Preetum Nakkiran, et al.
18

We introduce a new notion of generalization – Distributional Generalization – which roughly states that outputs of a classifier at train and test time are close *as distributions*, as opposed to close in just their average error. For example, if we mislabel 30 a ResNet trained to interpolation will in fact mislabel roughly 30 cats on the *test set* as well, while leaving other classes unaffected. This behavior is not captured by classical generalization, which would only consider the average error and not the distribution of errors over the input domain. Our formal conjectures, which are much more general than this example, characterize the form of distributional generalization that can be expected in terms of problem parameters: model architecture, training procedure, number of samples, and data distribution. We give empirical evidence for these conjectures across a variety of domains in machine learning, including neural networks, kernel machines, and decision trees. Our results thus advance our empirical understanding of interpolating classifiers.

READ FULL TEXT

page 4

page 5

page 14

page 15

page 21

page 31

page 34

page 36

research
06/04/2021

Out-of-Distribution Generalization in Kernel Regression

In real word applications, data generating process for training a machin...
research
04/07/2022

What You See is What You Get: Distributional Generalization for Algorithm Design in Deep Learning

We investigate and leverage a connection between Differential Privacy (D...
research
12/20/2022

Distributional Robustness Bounds Generalization Errors

Bayesian methods, distributionally robust optimization methods, and regu...
research
05/27/2019

Quantifying the generalization error in deep learning in terms of data distribution and neural network smoothness

The accuracy of deep learning, i.e., deep neural networks, can be charac...
research
05/13/2019

Affine Variational Autoencoders: An Efficient Approach for Improving Generalization and Robustness to Distribution Shift

In this study, we propose the Affine Variational Autoencoder (AVAE), a v...
research
02/25/2021

An Online Learning Approach to Interpolation and Extrapolation in Domain Generalization

A popular assumption for out-of-distribution generalization is that the ...
research
02/19/2020

Deep regularization and direct training of the inner layers of Neural Networks with Kernel Flows

We introduce a new regularization method for Artificial Neural Networks ...

Please sign up or login with your details

Forgot password? Click here to reset