Exact Phase Transitions in Deep Learning

05/25/2022
by   Liu Ziyin, et al.
0

This work reports deep-learning-unique first-order and second-order phase transitions, whose phenomenology closely follows that in statistical physics. In particular, we prove that the competition between prediction error and model complexity in the training loss leads to the second-order phase transition for nets with one hidden layer and the first-order phase transition for nets with more than one hidden layer. The proposed theory is directly relevant to the optimization of neural networks and points to an origin of the posterior collapse problem in Bayesian deep learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/31/2018

Convex Relaxations of Convolutional Neural Nets

We propose convex relaxations for convolutional neural nets with one hid...
research
01/07/2020

Phase Transitions for the Information Bottleneck in Representation Learning

In the Information Bottleneck (IB), when tuning the relative strength be...
research
03/29/2023

The Hidden-Manifold Hopfield Model and a learning phase transition

The Hopfield model has a long-standing tradition in statistical physics,...
research
07/18/2022

Hidden Progress in Deep Learning: SGD Learns Parities Near the Computational Limit

There is mounting empirical evidence of emergent phenomena in the capabi...
research
10/26/2022

Grokking phase transitions in learning local rules with gradient descent

We discuss two solvable grokking (generalisation beyond overfitting) mod...
research
02/26/2017

Criticality & Deep Learning I: Generally Weighted Nets

Motivated by the idea that criticality and universality of phase transit...
research
12/09/2022

Simulating first-order phase transition with hierarchical autoregressive networks

We apply the Hierarchical Autoregressive Neural (HAN) network sampling a...

Please sign up or login with your details

Forgot password? Click here to reset