Faster high-accuracy log-concave sampling via algorithmic warm starts

02/20/2023
by   Jason M. Altschuler, et al.
0

Understanding the complexity of sampling from a strongly log-concave and log-smooth distribution π on ℝ^d to high accuracy is a fundamental problem, both from a practical and theoretical standpoint. In practice, high-accuracy samplers such as the classical Metropolis-adjusted Langevin algorithm (MALA) remain the de facto gold standard; and in theory, via the proximal sampler reduction, it is understood that such samplers are key for sampling even beyond log-concavity (in particular, for distributions satisfying isoperimetric assumptions). In this work, we improve the dimension dependence of this sampling problem to Õ(d^1/2), whereas the previous best result for MALA was Õ(d). This closes the long line of work on the complexity of MALA, and moreover leads to state-of-the-art guarantees for high-accuracy sampling under strong log-concavity and beyond (thanks to the aforementioned reduction). Our starting point is that the complexity of MALA improves to Õ(d^1/2), but only under a warm start (an initialization with constant Rényi divergence w.r.t. π). Previous algorithms took much longer to find a warm start than to use it, and closing this gap has remained an important open problem in the field. Our main technical contribution settles this problem by establishing the first Õ(d^1/2) Rényi mixing rates for the discretized underdamped Langevin diffusion. For this, we develop new differential-privacy-inspired techniques based on Rényi divergences with Orlicz–Wasserstein shifts, which allow us to sidestep longstanding challenges for proving fast convergence of hypocoercive differential equations.

READ FULL TEXT

Please sign up or login with your details

Forgot password? Click here to reset