Nesterov's Accelerated Gradient and Momentum as approximations to Regularised Update Descent

07/07/2016
by   Aleksandar Botev, et al.
0

We present a unifying framework for adapting the update direction in gradient-based iterative optimization methods. As natural special cases we re-derive classical momentum and Nesterov's accelerated gradient method, lending a new intuitive interpretation to the latter algorithm. We show that a new algorithm, which we term Regularised Gradient Descent, can converge more quickly than either Nesterov's algorithm or the classical momentum algorithm.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/11/2019

Conformal Symplectic and Relativistic Optimization

Although momentum-based optimization methods have had a remarkable impac...
research
06/02/2019

Generalized Momentum-Based Methods: A Hamiltonian Perspective

We take a Hamiltonian-based perspective to generalize Nesterov's acceler...
research
06/04/2021

A Discrete Variational Derivation of Accelerated Methods in Optimization

Many of the new developments in machine learning are connected with grad...
research
11/24/2020

Shuffling Gradient-Based Methods with Momentum

We combine two advanced ideas widely used in optimization for machine le...
research
04/05/2022

Gradient Descent Bit-Flipping Decoding with Momentum

In this paper, we propose a Gradient Descent Bit-Flipping (GDBF) decodin...
research
02/12/2020

Average-case Acceleration Through Spectral Density Estimation

We develop a framework for designing optimal quadratic optimization meth...
research
03/02/2021

Categorical Foundations of Gradient-Based Learning

We propose a categorical foundation of gradient-based machine learning a...

Please sign up or login with your details

Forgot password? Click here to reset