Over-Parameterized Deep Neural Networks Have No Strict Local Minima For Any Continuous Activations

12/28/2018
by   Dawei Li, et al.
0

In this paper, we study the loss surface of the over-parameterized fully connected deep neural networks. We prove that for any continuous activation functions, the loss function has no bad strict local minimum, both in the regular sense and in the sense of sets. This result holds for any convex and continuous loss function, and the data samples are only required to be distinct in at least one dimension. Furthermore, we show that bad local minima do exist for a class of activation functions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/27/2018

On the loss landscape of a class of deep neural networks with no bad local valleys

We identify a class of over-parameterized deep neural networks with stan...
research
01/22/2019

On Connected Sublevel Sets in Deep Learning

We study sublevel sets of the loss function in training deep neural netw...
research
10/30/2017

The loss surface and expressivity of deep convolutional neural networks

We analyze the expressiveness and loss surface of practical deep convolu...
research
02/25/2021

Spurious Local Minima Are Common for Deep Neural Networks with Piecewise Linear Activations

In this paper, it is shown theoretically that spurious local minima are ...
research
02/23/2022

On the Omnipresence of Spurious Local Minima in Certain Neural Network Training Problems

We study the loss landscape of training problems for deep artificial neu...
research
11/04/2019

Sub-Optimal Local Minima Exist for Almost All Over-parameterized Neural Networks

Does over-parameterization eliminate sub-optimal local minima for neural...
research
04/26/2017

The loss surface of deep and wide neural networks

While the optimization problem behind deep neural networks is highly non...

Please sign up or login with your details

Forgot password? Click here to reset