Reparameterization Gradients through Acceptance-Rejection Sampling Algorithms

by   Christian A. Naesseth, et al.

Variational inference using the reparameterization trick has enabled large-scale approximate Bayesian inference in complex probabilistic models, leveraging stochastic optimization to sidestep intractable expectations. The reparameterization trick is applicable when we can simulate a random variable by applying a differentiable deterministic function on an auxiliary random variable whose distribution is fixed. For many distributions of interest (such as the gamma or Dirichlet), simulation of random variables relies on acceptance-rejection sampling. The discontinuity introduced by the accept-reject step means that standard reparameterization tricks are not applicable. We propose a new method that lets us leverage reparameterization gradients even when variables are outputs of a acceptance-rejection sampling algorithm. Our approach enables reparameterization on a larger class of variational distributions. In several studies of real and synthetic data, we show that the variance of the estimator of the gradient is significantly lower than other state-of-the-art methods. This leads to faster convergence of stochastic gradient variational inference.


page 1

page 2

page 3

page 4


Stochastic gradient variational Bayes for gamma approximating distributions

While stochastic variational inference is relatively well known for scal...

Local Expectation Gradients for Doubly Stochastic Variational Inference

We introduce local expectation gradients which is a general purpose stoc...

Generalized Transformation-based Gradient

The reparameterization trick has become one of the most useful tools in ...

Implicit Reparameterization Gradients

By providing a simple and efficient way of computing low-variance gradie...

Smoothed Gradients for Stochastic Variational Inference

Stochastic variational inference (SVI) lets us scale up Bayesian computa...

Differentiable Antithetic Sampling for Variance Reduction in Stochastic Variational Inference

Stochastic optimization techniques are standard in variational inference...

A Forest Mixture Bound for Block-Free Parallel Inference

Coordinate ascent variational inference is an important algorithm for in...