Criticality versus uniformity in deep neural networks

04/10/2023
by   Aleksandar Bukva, et al.
0

Deep feedforward networks initialized along the edge of chaos exhibit exponentially superior training ability as quantified by maximum trainable depth. In this work, we explore the effect of saturation of the tanh activation function along the edge of chaos. In particular, we determine the line of uniformity in phase space along which the post-activation distribution has maximum entropy. This line intersects the edge of chaos, and indicates the regime beyond which saturation of the activation function begins to impede training efficiency. Our results suggest that initialization along the edge of chaos is a necessary but not sufficient condition for optimal trainability.

READ FULL TEXT

page 9

page 11

research
02/19/2019

On the Impact of the Activation Function on Deep Neural Networks Training

The weight initialization and the activation function of deep neural net...
research
06/07/2023

Edge conductivity in PtSe_2 nanostructures

PtSe_2 is a promising 2D material for nanoelectromechanical sensing and ...
research
10/11/2019

The Expressivity and Training of Deep Neural Networks: toward the Edge of Chaos?

Expressivity is one of the most significant issues in assessing neural n...
research
10/16/2022

Stability of Accuracy for the Training of DNNs Via the Uniform Doubling Condition

We study the stability of accuracy for the training of deep neural netwo...
research
07/23/2020

Nonclosedness of the Set of Neural Networks in Sobolev Space

We examine the closedness of the set of realized neural networks of a fi...
research
06/06/2022

The Neural Covariance SDE: Shaped Infinite Depth-and-Width Networks at Initialization

The logit outputs of a feedforward neural network at initialization are ...
research
10/08/2020

Randomized Overdrive Neural Networks

By processing audio signals in the time-domain with randomly weighted te...

Please sign up or login with your details

Forgot password? Click here to reset