A Realistic Example in 2 Dimension that Gradient Descent Takes Exponential Time to Escape Saddle Points

08/17/2020
by   Shiliang Zuo, et al.
0

Gradient descent is a popular algorithm in optimization, and its performance in convex settings is mostly well understood. In non-convex settings, it has been shown that gradient descent is able to escape saddle points asymptotically and converge to local minimizers [Lee et. al. 2016]. Recent studies also show a perturbed version of gradient descent is enough to escape saddle points efficiently [Jin et. al. 2015, Ge et. al. 2017]. In this paper we show a negative result: gradient descent may take exponential time to escape saddle points, with non-pathological two dimensional functions. While our focus is theoretical, we also conduct experiments verifying our theoretical result. Through our analysis we demonstrate that stochasticity is essential to escape saddle points efficiently.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/29/2017

Gradient Descent Can Take Exponential Time to Escape Saddle Points

Although gradient descent (GD) almost always escapes saddle points asymp...
research
05/09/2020

Perturbed gradient descent with occupation time

This paper develops further the idea of perturbed gradient descent, by a...
research
01/21/2019

A Deterministic Approach to Avoid Saddle Points

Loss functions with a large number of saddle points are one of the main ...
research
01/09/2018

Convergence Analysis of Gradient Descent Algorithms with Proportional Updates

The rise of deep learning in recent years has brought with it increasing...
research
10/18/2019

First-Order Preconditioning via Hypergradient Descent

Standard gradient descent methods are susceptible to a range of issues t...
research
07/20/2022

A note on the variation of geometric functionals

Calculus of Variation combined with Differential Geometry as tools of mo...
research
02/13/2019

Stochastic Gradient Descent Escapes Saddle Points Efficiently

This paper considers the perturbed stochastic gradient descent algorithm...

Please sign up or login with your details

Forgot password? Click here to reset