The Natural Gradient by Analogy to Signal Whitening, and Recipes and Tricks for its Use

05/08/2012
by   Jascha Sohl-Dickstein, et al.
0

The natural gradient allows for more efficient gradient descent by removing dependencies and biases inherent in a function's parameterization. Several papers present the topic thoroughly and precisely. It remains a very difficult idea to get your head around however. The intent of this note is to provide simple intuition for the natural gradient and its use. We review how an ill conditioned parameter space can undermine learning, introduce the natural gradient by analogy to the more widely understood concept of signal whitening, and present tricks and specific prescriptions for applying the natural gradient to learning problems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/22/2021

Structured second-order methods via natural gradient descent

In this paper, we propose new structured second-order methods and struct...
research
11/22/2021

Depth Without the Magic: Inductive Bias of Natural Gradient Descent

In gradient descent, changing how we parametrize the model can lead to d...
research
06/29/2020

Natural Gradient for Combined Loss Using Wavelets

Natural gradients have been widely used in optimization of loss function...
research
01/19/2020

Dual Stochastic Natural Gradient Descent

Although theoretically appealing, Stochastic Natural Gradient Descent (S...
research
02/13/2022

Efficient Natural Gradient Descent Methods for Large-Scale Optimization Problems

We propose an efficient numerical method for computing natural gradient ...
research
03/24/2018

Natural Gradients in Practice: Non-Conjugate Variational Inference in Gaussian Process Models

The natural gradient method has been used effectively in conjugate Gauss...

Please sign up or login with your details

Forgot password? Click here to reset