Depth creates no more spurious local minima

01/28/2019
by   Li Zhang, et al.
0

We show that for any convex differentiable loss function, a deep linear network has no spurious local minima as long as it is true for the two layer case. When applied to the quadratic loss, our result immediately implies the powerful result in [Kawaguchi 2016] that there is no spurious local minima in deep linear networks. Further, with the recent work [Zhou and Liang 2018], we can remove all the assumptions in [Kawaguchi 2016]. Our proof is short and elementary. It builds on the recent work of [Laurent and von Brecht 2018] and uses a new rank one perturbation argument.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/05/2017

Deep linear neural networks with arbitrary loss: All local minima are global

We consider deep linear networks with arbitrary differentiable loss. We ...
research
06/28/2022

Diffeomorphic Registration using Sinkhorn Divergences

The diffeomorphic registration framework enables to define an optimal ma...
research
01/12/2019

Eliminating all bad Local Minima from Loss Landscapes without even adding an Extra Unit

Recent work has noted that all bad local minima can be removed from neur...
research
10/21/2018

Depth with Nonlinearity Creates No Bad Local Minima in ResNets

In this paper, we prove that depth with nonlinearity creates no bad loca...
research
02/16/2019

Making Convex Loss Functions Robust to Outliers using e-Exponentiated Transformation

In this paper, we propose a novel e-exponentiated transformation, 0.5< e...
research
11/29/2019

Barcodes as summary of objective function's topology

We apply the canonical forms (barcodes) of gradient Morse complexes to e...
research
08/12/2015

Inappropriate use of L-BFGS, Illustrated on frame field design

L-BFGS is a hill climbing method that is guarantied to converge only for...

Please sign up or login with your details

Forgot password? Click here to reset