A mean-field limit for certain deep neural networks

06/01/2019
by   Dyego Araújo, et al.
0

Understanding deep neural networks (DNNs) is a key challenge in the theory of machine learning, with potential applications to the many fields where DNNs have been successfully used. This article presents a scaling limit for a DNN being trained by stochastic gradient descent. Our networks have a fixed (but arbitrary) number L≥ 2 of inner layers; N≫ 1 neurons per layer; full connections between layers; and fixed weights (or "random features" that are not trained) near the input and output. Our results describe the evolution of the DNN during training in the limit when N→ +∞, which we relate to a mean field model of McKean-Vlasov type. Specifically, we show that network weights are approximated by certain "ideal particles" whose distribution and dependencies are described by the mean-field model. A key part of the proof is to show existence and uniqueness for our McKean-Vlasov problem, which does not seem to be amenable to existing theory. Our paper extends previous work on the L=1 case by Mei, Montanari and Nguyen; Rotskoff and Vanden-Eijnden; and Sirignano and Spiliopoulos. We also complement recent independent work on L>1 by Sirignano and Spiliopoulos (who consider a less natural scaling limit) and Nguyen (who nonrigorously derives similar results).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/11/2021

Global Convergence of Three-layer Neural Networks in the Mean Field Regime

In the mean field regime, neural networks are appropriately scaled so th...
research
10/04/2017

Mean-field theory of input dimensionality reduction in unsupervised deep neural networks

Deep neural networks as powerful tools are widely used in various domain...
research
10/09/2018

Information Geometry of Orthogonal Initializations and Training

Recently mean field theory has been successfully used to analyze propert...
research
03/24/2022

Extended critical regimes of deep neural networks

Deep neural networks (DNNs) have been successfully applied to many real-...
research
10/11/2020

Convergence to the fixed-node limit in deep variational Monte Carlo

Variational quantum Monte Carlo (QMC) is an ab-initio method for solving...
research
03/09/2023

Variational formulations of ODE-Net as a mean-field optimal control problem and existence results

This paper presents a mathematical analysis of ODE-Net, a continuum mode...
research
10/13/2019

Large Deviation Analysis of Function Sensitivity in Random Deep Neural Networks

Mean field theory has been successfully used to analyze deep neural netw...

Please sign up or login with your details

Forgot password? Click here to reset