Subsampling Error in Stochastic Gradient Langevin Diffusions

05/23/2023
by   Kexin Jin, et al.
0

The Stochastic Gradient Langevin Dynamics (SGLD) are popularly used to approximate Bayesian posterior distributions in statistical learning procedures with large-scale data. As opposed to many usual Markov chain Monte Carlo (MCMC) algorithms, SGLD is not stationary with respect to the posterior distribution; two sources of error appear: The first error is introduced by an Euler–Maruyama discretisation of a Langevin diffusion process, the second error comes from the data subsampling that enables its use in large-scale data settings. In this work, we consider an idealised version of SGLD to analyse the method's pure subsampling error that we then see as a best-case error for diffusion-based subsampling MCMC methods. Indeed, we introduce and study the Stochastic Gradient Langevin Diffusion (SGLDiff), a continuous-time Markov process that follows the Langevin diffusion corresponding to a data subset and switches this data subset after exponential waiting times. There, we show that the Wasserstein distance between the posterior and the limiting distribution of SGLDiff is bounded above by a fractional power of the mean waiting time. Importantly, this fractional power does not depend on the dimension of the state space. We bring our results into context with other analyses of SGLD.

READ FULL TEXT
research
06/05/2017

Stochastic Gradient Monomial Gamma Sampler

Recent advances in stochastic gradient techniques have made it possible ...
research
06/19/2018

Large-Scale Stochastic Sampling from the Probability Simplex

Stochastic gradient Markov chain Monte Carlo (SGMCMC) has become a popul...
research
12/23/2015

High-Order Stochastic Gradient Thermostats for Bayesian Learning of Deep Models

Learning in deep models using Bayesian methods has generated significant...
research
06/01/2023

Reconstructing Graph Diffusion History from a Single Snapshot

Diffusion on graphs is ubiquitous with numerous high-impact applications...
research
06/12/2018

Meta-Learning for Stochastic Gradient MCMC

Stochastic gradient Markov chain Monte Carlo (SG-MCMC) has become increa...
research
01/02/2020

Stochastic Gradient Langevin Dynamics on a Distributed Network

Langevin MCMC gradient optimization is a class of increasingly popular m...
research
11/01/2022

Bayesian Parameter Inference for Partially Observed SDEs driven by Fractional Brownian Motion

In this paper we consider Bayesian parameter inference for partially obs...

Please sign up or login with your details

Forgot password? Click here to reset