Metropolis-Hastings with Averaged Acceptance Ratios

12/29/2020 ∙ by Christophe Andrieu, et al. ∙ 0 ∙

Markov chain Monte Carlo (MCMC) methods to sample from a probability distribution π defined on a space (Θ,𝒯) consist of the simulation of realisations of Markov chains {θ_n,n≥1} of invariant distribution π and such that the distribution of θ_i converges to π as i→∞. In practice one is typically interested in the computation of expectations of functions, say f, with respect to π and it is also required that averages M^-1∑_n=1^Mf(θ_n) converge to the expectation of interest. The iterative nature of MCMC makes it difficult to develop generic methods to take advantage of parallel computing environments when interested in reducing time to convergence. While numerous approaches have been proposed to reduce the variance of ergodic averages, including averaging over independent realisations of {θ_n,n≥1} simulated on several computers, techniques to reduce the "burn-in" of MCMC are scarce. In this paper we explore a simple and generic approach to improve convergence to equilibrium of existing algorithms which rely on the Metropolis-Hastings (MH) update, the main building block of MCMC. The main idea is to use averages of the acceptance ratio w.r.t. multiple realisations of random variables involved, while preserving π as invariant distribution. The methodology requires limited change to existing code, is naturally suited to parallel computing and is shown on our examples to provide substantial performance improvements both in terms of convergence to equilibrium and variance of ergodic averages. In some scenarios gains are observed even on a serial machine.



There are no comments yet.


page 1

page 2

page 3

page 4

This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.