Collapse of Deep and Narrow Neural Nets

08/15/2018
by   Lu Lu, et al.
0

Recent theoretical work has demonstrated that deep neural networks have superior performance over shallow networks, but their training is more difficult, e.g., they suffer from the vanishing gradient problem. This problem can be typically resolved by the rectified linear unit (ReLU) activation. However, here we show that even for such activation, deep and narrow neural networks will converge to erroneous mean or median states of the target function depending on the loss with high probability. We demonstrate this collapse of deep and narrow neural networks both numerically and theoretically, and provide estimates of the probability of collapse. We also construct a diagram of a safe region of designing neural networks that avoid the collapse to erroneous states. Finally, we examine different ways of initialization and normalization that may avoid the collapse problem.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/21/2018

On the Selection of Initialization and Activation Function for Deep Neural Networks

The weight initialization and the activation function of deep neural net...
research
07/03/2018

On decision regions of narrow deep neural networks

We show that for neural network functions that have width less or equal ...
research
10/21/2022

When Expressivity Meets Trainability: Fewer than n Neurons Can Work

Modern neural networks are often quite wide, causing large memory and co...
research
09/13/2019

Shapley Interpretation and Activation in Neural Networks

We propose a novel Shapley value approach to help address neural network...
research
02/07/2022

Neural Tangent Kernel Analysis of Deep Narrow Neural Networks

The tremendous recent progress in analyzing the training dynamics of ove...
research
03/05/2018

How to Start Training: The Effect of Initialization and Architecture

We investigate the effects of initialization and architecture on the sta...
research
07/03/2019

Deep Personalized Re-targeting

Predicting booking probability and value at the traveler level plays a c...

Please sign up or login with your details

Forgot password? Click here to reset