On some variance reduction properties of the reparameterization trick

09/27/2018
by   Ming Xu, et al.
0

The so-called reparameterization trick is widely used in variational inference as it yields more accurate estimates of the gradient of the variational objective than alternative approaches such as the score function method. The resulting optimization converges much faster as the variance reduction offered by the reparameterization gradient is typically several orders of magnitude. There is overwhelming empirical evidence in the literature showing its success. However, there is relatively little research that explores why the reparameterization gradient is so effective. We explore this under two main simplifying assumptions. First, we assume that the variational approximation is the commonly used mean-field Gaussian density. Second, we assume that the log of the joint density of the model parameter vector and the data is a quadratic function that depends on the variational mean. These assumptions allow us to obtain tractable expressions for the marginal variances of the score function and reparameterization gradient estimators. We also derive lower bounds for the score function marginal variances through Rao-Blackwellization and prove that under our assumptions they are larger than those of the reparameterization trick. Finally, we apply the result of our idealized analysis to examples where the log-joint density is not quadratic, such as in a multinomial logistic regression and a Bayesian neural network with two layers.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/27/2018

Variance reduction properties of the reparameterization trick

The reparameterization trick is widely used in variational inference as ...
research
06/27/2012

Variational Bayesian Inference with Stochastic Search

Mean-field variational inference is a method for approximate Bayesian po...
research
11/19/2017

A note on quadratic approximations of logistic log-likelihoods

Quadratic approximations of logistic log-likelihoods are fundamental to ...
research
12/02/2019

Stochastic Variational Inference via Upper Bound

Stochastic variational inference (SVI) plays a key role in Bayesian deep...
research
10/19/2022

Second order stochastic gradient update for Cholesky factor in Gaussian variational approximation from Stein's Lemma

In stochastic variational inference, use of the reparametrization trick ...
research
03/18/2023

Practical and Matching Gradient Variance Bounds for Black-Box Variational Bayesian Inference

Understanding the gradient variance of black-box variational inference (...
research
01/26/2021

Generalized Doubly Reparameterized Gradient Estimators

Efficient low-variance gradient estimation enabled by the reparameteriza...

Please sign up or login with your details

Forgot password? Click here to reset