Amortized Variational Inference: When and Why?

07/20/2023
by   Charles C. Margossian, et al.
0

Amortized variational inference (A-VI) is a method for approximating the intractable posterior distributions that arise in probabilistic models. The defining feature of A-VI is that it learns a global inference function that maps each observation to its local latent variable's approximate posterior. This stands in contrast to the more classical factorized (or mean-field) variational inference (F-VI), which directly learns the parameters of the approximating distribution for each latent variable. In deep generative models, A-VI is used as a computational trick to speed up inference for local latent variables. In this paper, we study A-VI as a general alternative to F-VI for approximate posterior inference. A-VI cannot produce an approximation with a lower Kullback-Leibler divergence than F-VI's optimal solution, because the amortized family is a subset of the factorized family. Thus a central theoretical problem is to characterize when A-VI still attains F-VI's optimal solution. We derive conditions on both the model and the inference function under which A-VI can theoretically achieve F-VI's optimum. We show that for a broad class of hierarchical models, including deep generative models, it is possible to close the gap between A-VI and F-VI. Further, for an even broader class of models, we establish when and how to expand the domain of the inference function to make amortization a feasible strategy. Finally, we prove that for certain models – including hidden Markov models and Gaussian processes – A-VI cannot match F-VI's solution, no matter how expressive the inference function is. We also study A-VI empirically. On several examples, we corroborate our theoretical results and investigate the performance of A-VI when varying the complexity of the inference function. When the gap between A-VI and F-VI can be closed, we find that the required complexity of the function need not scale with the number of observations, and that A-VI often converges faster than F-VI.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/28/2015

Automatic Relevance Determination For Deep Generative Models

A recurring problem when building probabilistic latent variable models i...
research
11/20/2015

The Variational Gaussian Process

Variational inference is a powerful tool for approximate inference, and ...
research
03/18/2019

Approximating exponential family models (not single distributions) with a two-network architecture

Recently much attention has been paid to deep generative models, since t...
research
11/30/2022

Variational Laplace Autoencoders

Variational autoencoders employ an amortized inference model to approxim...
research
07/24/2018

Iterative Amortized Inference

Inference models are a key component in scaling variational inference to...
research
05/14/2019

Moment-Based Variational Inference for Markov Jump Processes

We propose moment-based variational inference as a flexible framework fo...
research
11/06/2018

A Novel Variational Family for Hidden Nonlinear Markov Models

Latent variable models have been widely applied for the analysis and vis...

Please sign up or login with your details

Forgot password? Click here to reset