Reparameterization Gradients through Acceptance-Rejection Sampling Algorithms

10/18/2016
by   Christian A. Naesseth, et al.
0

Variational inference using the reparameterization trick has enabled large-scale approximate Bayesian inference in complex probabilistic models, leveraging stochastic optimization to sidestep intractable expectations. The reparameterization trick is applicable when we can simulate a random variable by applying a differentiable deterministic function on an auxiliary random variable whose distribution is fixed. For many distributions of interest (such as the gamma or Dirichlet), simulation of random variables relies on acceptance-rejection sampling. The discontinuity introduced by the accept-reject step means that standard reparameterization tricks are not applicable. We propose a new method that lets us leverage reparameterization gradients even when variables are outputs of a acceptance-rejection sampling algorithm. Our approach enables reparameterization on a larger class of variational distributions. In several studies of real and synthetic data, we show that the variance of the estimator of the gradient is significantly lower than other state-of-the-art methods. This leads to faster convergence of stochastic gradient variational inference.

READ FULL TEXT

page 1

page 2

page 3

page 4

09/04/2015

Stochastic gradient variational Bayes for gamma approximating distributions

While stochastic variational inference is relatively well known for scal...
03/04/2015

Local Expectation Gradients for Doubly Stochastic Variational Inference

We introduce local expectation gradients which is a general purpose stoc...
11/06/2019

Generalized Transformation-based Gradient

The reparameterization trick has become one of the most useful tools in ...
05/22/2018

Implicit Reparameterization Gradients

By providing a simple and efficient way of computing low-variance gradie...
06/13/2014

Smoothed Gradients for Stochastic Variational Inference

Stochastic variational inference (SVI) lets us scale up Bayesian computa...
10/05/2018

Differentiable Antithetic Sampling for Variance Reduction in Stochastic Variational Inference

Stochastic optimization techniques are standard in variational inference...
05/17/2018

A Forest Mixture Bound for Block-Free Parallel Inference

Coordinate ascent variational inference is an important algorithm for in...