Mean field Variational Inference via Wasserstein Gradient Flow

07/17/2022
by   Rentian Yao, et al.
0

Variational inference (VI) provides an appealing alternative to traditional sampling-based approaches for implementing Bayesian inference due to its conceptual simplicity, statistical accuracy and computational scalability. However, common variational approximation schemes, such as the mean-field (MF) approximation, require certain conjugacy structure to facilitate efficient computation, which may add unnecessary restrictions to the viable prior distribution family and impose further constraints on the variational approximation family. In this work, we develop a general computational framework for implementing MF-VI via Wasserstein gradient flow (WGF), a gradient flow over the space of probability measures. When specialized to Bayesian latent variable models, we analyze the algorithmic convergence of an alternating minimization scheme based on a time-discretized WGF for implementing the MF approximation. In particular, the proposed algorithm resembles a distributional version of EM algorithm, consisting of an E-step of updating the latent variable variational distribution and an M-step of conducting steepest descent over the variational distribution of parameters. Our theoretical analysis relies on optimal transport theory and subdifferential calculus in the space of probability measures. We prove the exponential convergence of the time-discretized WGF for minimizing a generic objective functional given strict convexity along generalized geodesics. We also provide a new proof of the exponential contraction of the variational distribution obtained from the MF approximation by using the fixed-point equation of the time-discretized WGF. We apply our method and theory to two classic Bayesian latent variable models, the Gaussian mixture model and the mixture of regression model. Numerical experiments are also conducted to compliment the theoretical findings under these two models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/20/2022

On Representations of Mean-Field Variational Inference

The mean field variational inference (MFVI) formulation restricts the ge...
research
06/01/2023

On the Convergence of Coordinate Ascent Variational Inference

As a computational alternative to Markov chain Monte Carlo approaches, v...
research
11/15/2022

Regularized Stein Variational Gradient Flow

The Stein Variational Gradient Descent (SVGD) algorithm is an determinis...
research
12/10/2019

Frequentist Consistency of Generalized Variational Inference

This paper investigates Frequentist consistency properties of the poster...
research
10/07/2020

Learning Deep-Latent Hierarchies by Stacking Wasserstein Autoencoders

Probabilistic models with hierarchical-latent-variable structures provid...
research
03/07/2018

The Ising distribution as a latent variable model

We show that the Ising distribution can be treated as a latent variable ...
research
05/20/2020

Infinite-dimensional gradient-based descent for alpha-divergence minimisation

This paper introduces the (α, Γ)-descent, an iterative algorithm which o...

Please sign up or login with your details

Forgot password? Click here to reset