The shifted ODE method for underdamped Langevin MCMC
In this paper, we consider the underdamped Langevin diffusion (ULD) and propose a numerical approximation using its associated ordinary differential equation (ODE). When used as a Markov Chain Monte Carlo (MCMC) algorithm, we show that the ODE approximation achieves a 2-Wasserstein error of ε in 𝒪(d^1/3/ε^2/3) steps under the standard smoothness and strong convexity assumptions on the target distribution. This matches the complexity of the randomized midpoint method proposed by Shen and Lee [NeurIPS 2019] which was shown to be order optimal by Cao, Lu and Wang. However, the main feature of the proposed numerical method is that it can utilize additional smoothness of the target log-density f. More concretely, we show that the ODE approximation achieves a 2-Wasserstein error of ε in 𝒪(d^2/5/ε^2/5) and 𝒪(√(d)/ε^1/3) steps when Lipschitz continuity is assumed for the Hessian and third derivative of f. By discretizing this ODE using a third order Runge-Kutta method, we can obtain a practical MCMC method that uses just two additional gradient evaluations per step. In our experiment, where the target comes from a logistic regression, this method shows faster convergence compared to other unadjusted Langevin MCMC algorithms.
READ FULL TEXT