Explicit natural gradient updates for Cholesky factor in Gaussian variational approximation

09/01/2021
by   Linda S. L. Tan, et al.
0

Stochastic gradient methods have enabled variational inference for high-dimensional models and large data. However, the steepest ascent direction in the parameter space of a statistical model is given not by the commonly used Euclidean gradient, but the natural gradient which premultiplies the Euclidean gradient by the inverted Fisher information matrix. Use of natural gradients can improve convergence significantly, but inverting the Fisher information matrix is daunting in high-dimensions. In Gaussian variational approximation, natural gradient updates of the natural parameters (expressed in terms of the mean and precision matrix) of the Gaussian distribution can be derived analytically, but do not ensure the precision matrix remains positive definite. To tackle this issue, we consider Cholesky decomposition of the covariance or precision matrix and derive explicit natural gradient updates of the Cholesky factor by finding the inverse of the Fisher information matrix analytically. Natural gradient updates of the Cholesky factor as compared to natural parameters, depend only on the first instead of the second derivative of the log posterior density and reduces computational cost. Sparsity constraints incorporating posterior independence structure can be imposed by fixing relevant entries in the Cholesky factor to zero.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/19/2022

Second order stochastic gradient update for Cholesky factor in Gaussian variational approximation from Stein's Lemma

In stochastic variational inference, use of the reparametrization trick ...
research
05/18/2016

Gaussian variational approximation with sparse precision matrices

We consider the problem of learning a Gaussian variational approximation...
research
03/07/2019

The Variational Predictive Natural Gradient

Variational inference transforms posterior inference into parametric opt...
research
02/27/2023

Natural Gradient Hybrid Variational Inference with Application to Deep Mixed Models

Stochastic models with global parameters θ and latent variables z are co...
research
12/04/2017

Natural Langevin Dynamics for Neural Networks

One way to avoid overfitting in machine learning is to use model paramet...
research
06/14/2021

NG+ : A Multi-Step Matrix-Product Natural Gradient Method for Deep Learning

In this paper, a novel second-order method called NG+ is proposed. By fo...
research
02/11/2019

Manifold Optimisation Assisted Gaussian Variational Approximation

Variational approximation methods are a way to approximate the posterior...

Please sign up or login with your details

Forgot password? Click here to reset