Predicting the Generalization Gap in Deep Networks with Margin Distributions

09/28/2018
by   Yiding Jiang, et al.
0

As shown in recent research, deep neural networks can perfectly fit randomly labeled data, but with very poor accuracy on held out data. This phenomenon indicates that loss functions such as cross-entropy are not a reliable indicator of generalization. This leads to the crucial question of how generalization gap should be predicted from the training data and network parameters. In this paper, we propose such a measure, and conduct extensive empirical studies on how well it can predict the generalization gap. Our measure is based on the concept of margin distribution, which are the distances of training points to the decision boundary. We find that it is necessary to use margin distributions at multiple layers of a deep network. On the CIFAR-10 and the CIFAR-100 datasets, our proposed measure correlates very strongly with the generalization gap. In addition, we find the following other factors to be of importance: normalizing margin values for scale independence, using characterizations of margin distribution rather than just the margin (closest distance to decision boundary), and working in log space instead of linear space (effectively using a product of margins rather than a sum). Our measure can be easily applied to feedforward deep networks with any architecture and may point towards new training loss functions that could enable better generalization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/27/2018

Optimal Margin Distribution Network

Recent research about margin theory has proved that maximizing the minim...
research
03/15/2018

Large Margin Deep Networks for Classification

We present a formulation of deep learning that aims at producing a large...
research
08/29/2023

Input margins can predict generalization too

Understanding generalization in deep neural networks is an active area o...
research
10/12/2018

On the Margin Theory of Feedforward Neural Networks

Past works have shown that, somewhat surprisingly, over-parametrization ...
research
06/11/2022

Learning Imbalanced Datasets with Maximum Margin Loss

A learning algorithm referred to as Maximum Margin (MM) is proposed for ...
research
03/09/2021

Knowledge Evolution in Neural Networks

Deep learning relies on the availability of a large corpus of data (labe...
research
10/09/2019

Improved Sample Complexities for Deep Networks and Robust Classification via an All-Layer Margin

For linear classifiers, the relationship between (normalized) output mar...

Please sign up or login with your details

Forgot password? Click here to reset