First-order and second-order variants of the gradient descent: a unified framework

10/18/2018
by   Thomas Pierrot, et al.
0

In this paper, we provide an overview of first-order and second-order variants of the gradient descent methods commonly used in machine learning. We propose a general framework in which 6 of these methods can be interpreted as different instances of the same approach. These methods are the vanilla gradient descent, the classical and generalized Gauss-Newton methods, the natural gradient descent method, the gradient covariance matrix approach, and Newton's method. Besides interpreting these methods within a single framework, we explain their specificities and show under which conditions some of them coincide.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/22/2021

Structured second-order methods via natural gradient descent

In this paper, we propose new structured second-order methods and struct...
research
09/26/2018

Learning Preconditioners on Lie Groups

We study two types of preconditioners and preconditioned stochastic grad...
research
08/26/2020

Gravilon: Applications of a New Gradient Descent Method to Machine Learning

Gradient descent algorithms have been used in countless applications sin...
research
11/08/2021

Inertial Newton Algorithms Avoiding Strict Saddle Points

We study the asymptotic behavior of second-order algorithms mixing Newto...
research
01/16/2013

Revisiting Natural Gradient for Deep Networks

We evaluate natural gradient, an algorithm originally proposed in Amari ...
research
06/17/2020

A block coordinate descent optimizer for classification problems exploiting convexity

Second-order optimizers hold intriguing potential for deep learning, but...
research
08/16/2021

Adaptive Gradient Descent Methods for Computing Implied Volatility

In this paper, a new numerical method based on adaptive gradient descent...

Please sign up or login with your details

Forgot password? Click here to reset