Slope and generalization properties of neural networks

07/03/2021
by   Anton Johansson, et al.
0

Neural networks are very successful tools in for example advanced classification. From a statistical point of view, fitting a neural network may be seen as a kind of regression, where we seek a function from the input space to a space of classification probabilities that follows the "general" shape of the data, but avoids overfitting by avoiding memorization of individual data points. In statistics, this can be done by controlling the geometric complexity of the regression function. We propose to do something similar when fitting neural networks by controlling the slope of the network. After defining the slope and discussing some of its theoretical properties, we go on to show empirically in examples, using ReLU networks, that the distribution of the slope of a well-trained neural network classifier is generally independent of the width of the layers in a fully connected network, and that the mean of the distribution only has a weak dependence on the model architecture in general. The slope is of similar size throughout the relevant volume, and varies smoothly. It also behaves as predicted in rescaling examples. We discuss possible applications of the slope concept, such as using it as a part of the loss function or stopping criterion during network training, or ranking data sets in terms of their complexity.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/26/2019

Emergent Structures and Lifetime Structure Evolution in Artificial Neural Networks

Motivated by the flexibility of biological neural networks whose connect...
research
12/15/2021

Robust Neural Network Classification via Double Regularization

The presence of mislabeled observations in data is a notoriously challen...
research
01/28/2019

Stiffness: A New Perspective on Generalization in Neural Networks

We investigate neural network training and generalization using the conc...
research
07/05/2022

Correlation between entropy and generalizability in a neural network

Although neural networks can solve very complex machine-learning problem...
research
02/14/2019

Generalisation in fully-connected neural networks for time series forecasting

In this paper we study the generalisation capabilities of fully-connecte...
research
10/02/2018

GINN: Geometric Illustration of Neural Networks

This informal technical report details the geometric illustration of dec...
research
04/02/2019

On Geometric Structure of Activation Spaces in Neural Networks

In this paper, we investigate the geometric structure of activation spac...

Please sign up or login with your details

Forgot password? Click here to reset