Does Hamiltonian Monte Carlo mix faster than a random walk on multimodal densities?

08/09/2018
by   Oren Mangoubi, et al.
0

Hamiltonian Monte Carlo (HMC) is a very popular and generic collection of Markov chain Monte Carlo (MCMC) algorithms. One explanation for the popularity of HMC algorithms is their excellent performance as the dimension d of the target becomes large: under conditions that are satisfied for many common statistical models, optimally-tuned HMC algorithms have a running time that scales like d^0.25. In stark contrast, the running time of the usual Random-Walk Metropolis (RWM) algorithm, optimally tuned, scales like d. This superior scaling of the HMC algorithm with dimension is attributed to the fact that it, unlike RWM, incorporates the gradient information in the proposal distribution. In this paper, we investigate a different scaling question: does HMC beat RWM for highly multimodal targets? We find that the answer is often no. We compute the spectral gaps for both the algorithms for a specific class of multimodal target densities, and show that they are identical. The key reason is that, within one mode, the gradient is effectively ignorant about other modes, thus negating the advantage the HMC algorithm enjoys in unimodal targets. We also give heuristic arguments suggesting that the above observation may hold quite generally. Our main tool for answering this question is a novel simple formula for the conductance of HMC using Liouville's theorem. This result allows us to compute the spectral gap of HMC algorithms, for both the classical HMC with isotropic momentum and the recent Riemannian HMC, for multimodal targets.

READ FULL TEXT
research
01/14/2020

Hamiltonian Monte Carlo Swindles

Hamiltonian Monte Carlo (HMC) is a powerful Markov chain Monte Carlo (MC...
research
12/23/2014

Particle Metropolis-adjusted Langevin algorithms

This paper proposes a new sampling scheme based on Langevin dynamics tha...
research
05/17/2020

Hamiltonian Assisted Metropolis Sampling

Various Markov chain Monte Carlo (MCMC) methods are studied to improve u...
research
10/21/2019

Counterexamples for optimal scaling of Metropolis-Hastings chains with rough target densities

For sufficiently smooth targets of product form it is known that the var...
research
08/09/2018

Simple Conditions for Metastability of Continuous Markov Chains

A family {Q_β}_β≥ 0 of Markov chains is said to exhibit metastable mixin...
research
08/30/2019

On the robustness of gradient-based MCMC algorithms

We analyse the tension between robustness and efficiency for Markov chai...
research
08/14/2018

Weight-Preserving Simulated Tempering

Simulated tempering is popular method of allowing MCMC algorithms to mov...

Please sign up or login with your details

Forgot password? Click here to reset