Fast yet Simple Natural-Gradient Descent for Variational Inference in Complex Models

07/12/2018
by   Mohammad Emtiyaz Khan, et al.
0

Bayesian inference plays an important role in advancing machine learning, but faces computational challenges when applied to complex models such as deep neural networks. Variational inference circumvents these challenges by formulating Bayesian inference as an optimization problem and solving it using gradient-based optimization. In this paper, we argue in favor of natural-gradient approaches which, unlike their gradient-based counterparts, can improve convergence by exploiting the information geometry of the solutions. We show how to derive fast yet simple natural-gradient updates by using a duality associated with exponential-family distributions. An attractive feature of these methods is that, by using natural-gradients, they are able to extract accurate local approximations for individual model components. We summarize recent results for Bayesian deep learning showing the superiority of natural-gradient approaches over their gradient counterparts.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/07/2019

Fast and Simple Natural-Gradient Variational Inference with Mixture of Exponential-family Approximations

Natural-gradient methods enable fast and simple algorithms for variation...
research
07/22/2021

Neural Variational Gradient Descent

Particle-based approximate Bayesian inference approaches such as Stein V...
research
11/21/2022

Bayesian Learning for Neural Networks: an algorithmic survey

The last decade witnessed a growing interest in Bayesian learning. Yet, ...
research
11/12/2021

Predictive coding, precision and natural gradients

There is an increasing convergence between biologically plausible comput...
research
04/13/2021

Stein variational gradient descent with local approximations

Bayesian computation plays an important role in modern machine learning ...
research
06/12/2020

Approximate Inference for Spectral Mixture Kernel

A spectral mixture (SM) kernel is a flexible kernel used to model any st...
research
07/08/2021

Analytically Tractable Hidden-States Inference in Bayesian Neural Networks

With few exceptions, neural networks have been relying on backpropagatio...

Please sign up or login with your details

Forgot password? Click here to reset