Why Neural Networks Work

11/26/2022
by   Sayandev Mukherjee, et al.
0

We argue that many properties of fully-connected feedforward neural networks (FCNNs), also called multi-layer perceptrons (MLPs), are explainable from the analysis of a single pair of operations, namely a random projection into a higher-dimensional space than the input, followed by a sparsification operation. For convenience, we call this pair of successive operations expand-and-sparsify following the terminology of Dasgupta. We show how expand-and-sparsify can explain the observed phenomena that have been discussed in the literature, such as the so-called Lottery Ticket Hypothesis, the surprisingly good performance of randomly-initialized untrained neural networks, the efficacy of Dropout in training and most importantly, the mysterious generalization ability of overparameterized models, first highlighted by Zhang et al. and subsequently identified even in non-neural network models by Belkin et al.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/28/2019

Tensor Programs I: Wide Feedforward or Recurrent Neural Networks of Any Architecture are Gaussian Processes

Wide neural networks with random weights and biases are Gaussian process...
research
02/04/2020

A Deep Conditioning Treatment of Neural Networks

We study the role of depth in training randomly initialized overparamete...
research
02/19/2019

On the Impact of the Activation Function on Deep Neural Networks Training

The weight initialization and the activation function of deep neural net...
research
06/11/2019

Neural network identifiability for a family of sigmoidal nonlinearities

This paper addresses the following question of neural network identifiab...
research
06/08/2022

Neural Collapse: A Review on Modelling Principles and Generalization

With a recent observation of the "Neural Collapse (NC)" phenomena by Pap...
research
10/06/2003

On Interference of Signals and Generalization in Feedforward Neural Networks

This paper studies how the generalization ability of neurons can be affe...
research
02/24/2021

Abelian Neural Networks

We study the problem of modeling a binary operation that satisfies some ...

Please sign up or login with your details

Forgot password? Click here to reset