Eigenvalue Corrected Noisy Natural Gradient

11/30/2018
by   Juhan Bae, et al.
5

Variational Bayesian neural networks combine the flexibility of deep learning with Bayesian uncertainty estimation. However, inference procedures for flexible variational posteriors are computationally expensive. A recently proposed method, noisy natural gradient, is a surprisingly simple method to fit expressive posteriors by adding weight noise to regular natural gradient updates. Noisy K-FAC is an instance of noisy natural gradient that fits a matrix-variate Gaussian posterior with minor changes to ordinary K-FAC. Nevertheless, a matrix-variate Gaussian posterior does not capture an accurate diagonal variance. In this work, we extend on noisy K-FAC to obtain a more flexible posterior distribution called eigenvalue corrected matrix-variate Gaussian. The proposed method computes the full diagonal re-scaling factor in Kronecker-factored eigenbasis. Empirically, our approach consistently outperforms existing algorithms (e.g., noisy K-FAC) on regression and classification tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/06/2017

Noisy Natural Gradient as Variational Inference

Combining the flexibility of deep learning with Bayesian uncertainty est...
research
11/27/2020

Eigenvalue-corrected Natural Gradient Based on a New Approximation

Using second-order optimization methods for training deep neural network...
research
11/11/2018

SLANG: Fast Structured Covariance Approximations for Bayesian Deep Learning with Natural Gradient

Uncertainty estimation in large deep-learning models is a computationall...
research
02/27/2023

Natural Gradient Hybrid Variational Inference with Application to Deep Mixed Models

Stochastic models with global parameters θ and latent variables z are co...
research
06/10/2021

Quantum Natural Gradient for Variational Bayes

Variational Bayes (VB) is a critical method in machine learning and stat...
research
02/07/2019

A Simple Baseline for Bayesian Uncertainty in Deep Learning

We propose SWA-Gaussian (SWAG), a simple, scalable, and general purpose ...
research
04/20/2020

Tractable Approximate Gaussian Inference for Bayesian Neural Networks

In this paper, we propose an analytical method allowing for tractable ap...

Please sign up or login with your details

Forgot password? Click here to reset