Rao-Blackwellized Stochastic Gradients for Discrete Distributions

10/10/2018
by   Runjing Liu, et al.
0

We wish to compute the gradient of an expectation over a finite or countably infinite sample space having K ≤∞ categories. When K is indeed infinite, or finite but very large, the relevant summation is intractable. Accordingly, various stochastic gradient estimators have been proposed. In this paper, we describe a technique that can be applied to reduce the variance of any such estimator, without changing its bias---in particular, unbiasedness is retained. We show that our technique is an instance of Rao-Blackwellization, and we demonstrate the improvement it yields in empirical studies on both synthetic and real-world data.

READ FULL TEXT
research
03/31/2019

Perturbative estimation of stochastic gradients

In this paper we introduce a family of stochastic gradient estimation te...
research
05/29/2022

Stochastic Zeroth Order Gradient and Hessian Estimators: Variance Reduction and Refined Bias Bounds

We study stochastic zeroth order gradient and Hessian estimators for rea...
research
02/19/2022

Gradient Estimation with Discrete Stein Operators

Gradient estimation – approximating the gradient of an expectation with ...
research
09/28/2021

An Accelerated Stochastic Gradient for Canonical Polyadic Decomposition

We consider the problem of structured canonical polyadic decomposition. ...
research
02/20/2020

Improving Sampling Accuracy of Stochastic Gradient MCMC Methods via Non-uniform Subsampling of Gradients

Common Stochastic Gradient MCMC methods approximate gradients by stochas...
research
09/29/2018

Improved Gradient-Based Optimization Over Discrete Distributions

In many applications we seek to maximize an expectation with respect to ...
research
10/05/2019

Straight-Through Estimator as Projected Wasserstein Gradient Flow

The Straight-Through (ST) estimator is a widely used technique for back-...

Please sign up or login with your details

Forgot password? Click here to reset