Stochastic optimization with momentum: convergence, fluctuations, and traps avoidance

12/07/2020
by   A. Barakat, et al.
0

In this paper, a general stochastic optimization procedure is studied, unifying several variants of the stochastic gradient descent such as, among others, the stochastic heavy ball method, the Stochastic Nesterov Accelerated Gradient algorithm (S-NAG), and the widely used Adam algorithm. The algorithm is seen as a noisy Euler discretization of a non-autonomous ordinary differential equation, recently introduced by Belotto da Silva and Gazeau, which is analyzed in depth. Assuming that the objective function is non-convex and differentiable, the stability and the almost sure convergence of the iterates to the set of critical points are established. A noteworthy special case is the convergence proof of S-NAG in a non-convex setting. Under some assumptions, the convergence rate is provided under the form of a Central Limit Theorem. Finally, the non-convergence of the algorithm to undesired critical points, such as local maxima or saddle points, is established. Here, the main ingredient is a new avoidance of traps result for non-autonomous settings, which is of independent interest.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/10/2018

On the Convergence of Weighted AdaGrad with Momentum for Training Deep Neural Networks

Adaptive stochastic gradient descent methods, such as AdaGrad, RMSProp, ...
research
10/04/2018

Convergence of the ADAM algorithm from a Dynamical System Viewpoint

Adam is a popular variant of the stochastic gradient descent for finding...
research
10/11/2017

Stochastic Gradient Descent in Continuous Time: A Central Limit Theorem

Stochastic gradient descent in continuous time (SGDCT) provides a comput...
research
05/13/2018

Dyna: A Method of Momentum for Stochastic Optimization

An algorithm is presented for momentum gradient descent optimization bas...
research
04/18/2014

iPiano: Inertial Proximal Algorithm for Non-Convex Optimization

In this paper we study an algorithm for solving a minimization problem c...
research
05/03/2022

Convergence of Stochastic Approximation via Martingale and Converse Lyapunov Methods

This paper is dedicated to Prof. Eduardo Sontag on the occasion of his s...
research
06/11/2018

Swarming for Faster Convergence in Stochastic Optimization

We study a distributed framework for stochastic optimization which is in...

Please sign up or login with your details

Forgot password? Click here to reset