Learning Preconditioners on Lie Groups

09/26/2018
by   Xi-Lin Li, et al.
0

We study two types of preconditioners and preconditioned stochastic gradient descent (SGD) methods in a unified framework. We call the first one the Newton type due to its close relationship to Newton method, and the second one the Fisher type as its preconditioner is closely related to the inverse of Fisher information matrix. Both preconditioners can be derived from one framework, and efficiently learned on any matrix Lie groups designated by the user using natural or relative gradient descent. Many existing preconditioners and methods are special cases of either the Newton type or the Fisher type ones. Experimental results on relatively large scale machine learning problems are reported for performance study.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/18/2018

First-order and second-order variants of the gradient descent: a unified framework

In this paper, we provide an overview of first-order and second-order va...
research
05/18/2023

Modified Gauss-Newton Algorithms under Noise

Gauss-Newton methods and their stochastic version have been widely used ...
research
02/15/2021

Tractable structured natural gradient descent using local parameterizations

Natural-gradient descent on structured parameter spaces (e.g., low-rank ...
research
12/03/2014

New insights and perspectives on the natural gradient method

Natural gradient descent is an optimization method traditionally motivat...
research
05/29/2019

Limitations of the Empirical Fisher Approximation

Natural gradient descent, which preconditions a gradient descent update ...
research
07/01/2015

Natural Neural Networks

We introduce Natural Neural Networks, a novel family of algorithms that ...
research
07/13/2022

Communication-efficient Distributed Newton-like Optimization with Gradients and M-estimators

In modern data science, it is common that large-scale data are stored an...

Please sign up or login with your details

Forgot password? Click here to reset