Phase Diagram of Initial Condensation for Two-layer Neural Networks

03/12/2023
by   Zhengan Chen, et al.
0

The phenomenon of distinct behaviors exhibited by neural networks under varying scales of initialization remains an enigma in deep learning research. In this paper, based on the earlier work by Luo et al. <cit.>, we present a phase diagram of initial condensation for two-layer neural networks. Condensation is a phenomenon wherein the weight vectors of neural networks concentrate on isolated orientations during the training process, and it is a feature in non-linear learning process that enables neural networks to possess better generalization abilities. Our phase diagram serves to provide a comprehensive understanding of the dynamical regimes of neural networks and their dependence on the choice of hyperparameters related to initialization. Furthermore, we demonstrate in detail the underlying mechanisms by which small initialization leads to condensation at the initial training stage.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/24/2022

Empirical Phase Diagram for Three-layer Neural Networks with Infinite Width

Substantial work indicates that the dynamics of neural networks (NNs) is...
research
05/17/2023

Understanding the Initial Condensation of Convolutional Neural Networks

Previous research has shown that fully-connected networks with small ini...
research
12/02/2021

Trap of Feature Diversity in the Learning of MLPs

In this paper, we discover a two-phase phenomenon in the learning of mul...
research
05/31/2021

Persistent Homology Captures the Generalization of Neural Networks Without A Validation Set

The training of neural networks is usually monitored with a validation (...
research
05/25/2021

Towards Understanding the Condensation of Two-layer Neural Networks at Initial Training

It is important to study what implicit regularization is imposed on the ...
research
07/01/2019

On Symmetry and Initialization for Neural Networks

This work provides an additional step in the theoretical understanding o...
research
04/21/2021

Deep limits and cut-off phenomena for neural networks

We consider dynamical and geometrical aspects of deep learning. For many...

Please sign up or login with your details

Forgot password? Click here to reset