A New Framework for Variance-Reduced Hamiltonian Monte Carlo

02/09/2021
by   Zhengmian Hu, et al.
0

We propose a new framework of variance-reduced Hamiltonian Monte Carlo (HMC) methods for sampling from an L-smooth and m-strongly log-concave distribution, based on a unified formulation of biased and unbiased variance reduction methods. We study the convergence properties for HMC with gradient estimators which satisfy the Mean-Squared-Error-Bias (MSEB) property. We show that the unbiased gradient estimators, including SAGA and SVRG, based HMC methods achieve highest gradient efficiency with small batch size under high precision regime, and require Õ(N + κ^2 d^1/2ε^-1 + N^2/3κ^4/3 d^1/3ε^-2/3 ) gradient complexity to achieve ϵ-accuracy in 2-Wasserstein distance. Moreover, our HMC methods with biased gradient estimators, such as SARAH and SARGE, require Õ(N+√(N)κ^2 d^1/2ε^-1) gradient complexity, which has the same dependency on condition number κ and dimension d as full gradient method, but improves the dependency of sample size N for a factor of N^1/2. Experimental results on both synthetic and real-world benchmark data show that our new framework significantly outperforms the full gradient and stochastic gradient HMC approaches. The earliest version of this paper was submitted to ICML 2020 with three weak accept but was not finally accepted.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2018

Stochastic Variance-Reduced Hamilton Monte Carlo Methods

We propose a fast stochastic Hamilton Monte Carlo (HMC) method, for samp...
research
03/29/2018

Stochastic Gradient Hamiltonian Monte Carlo with Variance Reduction for Bayesian Inference

Gradient-based Monte Carlo sampling algorithms, like Langevin dynamics a...
research
02/13/2020

Nonasymptotic analysis of Stochastic Gradient Hamiltonian Monte Carlo under local conditions for nonconvex optimization

We provide a nonasymptotic analysis of the convergence of the stochastic...
research
02/15/2018

On the Theory of Variance Reduction for Stochastic Gradient Monte Carlo

We provide convergence guarantees in Wasserstein distance for a variety ...
research
11/29/2022

Penalized Langevin and Hamiltonian Monte Carlo Algorithms for Constrained Sampling

We consider the constrained sampling problem where the goal is to sample...
research
06/17/2021

Stochastic Bias-Reduced Gradient Methods

We develop a new primitive for stochastic optimization: a low-bias, low-...
research
02/24/2018

Dimensionally Tight Bounds for Second-Order Hamiltonian Monte Carlo

Hamiltonian Monte Carlo (HMC) is a widely deployed method to sample from...

Please sign up or login with your details

Forgot password? Click here to reset