Depth Creates No Bad Local Minima

02/27/2017
by   Haihao Lu, et al.
0

In deep learning, depth, as well as nonlinearity, create non-convex loss surfaces. Then, does depth alone create bad local minima? In this paper, we prove that without nonlinearity, depth alone does not create bad local minima, although it induces non-convex loss surface. Using this insight, we greatly simplify a recently proposed proof to show that all of the local minima of feedforward deep linear neural networks are global minima. Our theoretical results generalize previous results with fewer assumptions, and this analysis provides a method to show similar results beyond square loss in deep linear models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/21/2018

Depth with Nonlinearity Creates No Bad Local Minima in ResNets

In this paper, we prove that depth with nonlinearity creates no bad loca...
research
09/12/2016

Non-square matrix sensing without spurious local minima via the Burer-Monteiro approach

We consider the non-square matrix sensing problem, under restricted isom...
research
11/19/2016

Local minima in training of neural networks

There has been a lot of recent interest in trying to characterize the er...
research
10/17/2018

The loss surface of deep linear networks viewed through the algebraic geometry lens

By using the viewpoint of modern computational algebraic geometry, we ex...
research
11/11/2021

Towards Theoretical Understanding of Flexible Transmitter Networks via Approximation and Local Minima

Flexible Transmitter Network (FTNet) is a recently proposed bio-plausibl...
research
07/20/2021

Learn2Hop: Learned Optimization on Rough Landscapes

Optimization of non-convex loss surfaces containing many local minima re...
research
03/18/2020

On the Distribution of Minima in Intrinsic-Metric Rotation Averaging

Rotation Averaging is a non-convex optimization problem that determines ...

Please sign up or login with your details

Forgot password? Click here to reset