Structure and Gradient Dynamics Near Global Minima of Two-layer Neural Networks

09/01/2023
by   Leyang Zhang, et al.
0

Under mild assumptions, we investigate the structure of loss landscape of two-layer neural networks near global minima, determine the set of parameters which give perfect generalization, and fully characterize the gradient flows around it. With novel techniques, our work uncovers some simple aspects of the complicated loss landscape and reveals how model, target function, samples and initialization affect the training dynamics differently. Based on these results, we also explain why (overparametrized) neural networks could generalize well.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/30/2017

Towards Understanding Generalization of Deep Learning: Perspective of Loss Landscapes

It is widely observed that deep learning models with learned parameters ...
research
12/31/2019

No Spurious Local Minima in Deep Quadratic Networks

Despite their practical success, a theoretical understanding of the loss...
research
04/09/2022

FuNNscope: Visual microscope for interactively exploring the loss landscape of fully connected neural networks

Despite their effective use in various fields, many aspects of neural ne...
research
04/22/2014

Attractor Metadynamics in Adapting Neural Networks

Slow adaption processes, like synaptic and intrinsic plasticity, abound ...
research
02/01/2019

Passed & Spurious: analysing descent algorithms and local minima in spiked matrix-tensor model

In this work we analyse quantitatively the interplay between the loss la...
research
06/14/2021

Extracting Global Dynamics of Loss Landscape in Deep Learning Models

Deep learning models evolve through training to learn the manifold in wh...
research
11/01/2017

Learning One-hidden-layer Neural Networks with Landscape Design

We consider the problem of learning a one-hidden-layer neural network: w...

Please sign up or login with your details

Forgot password? Click here to reset