On the Convergence of the ELBO to Entropy Sums

09/07/2022
by   Jörg Lücke, et al.
0

The variational lower bound (a.k.a. ELBO or free energy) is the central objective for many learning algorithms including algorithms for deep unsupervised learning. Learning algorithms change model parameters such that the variational lower bound increases, and until the parameters are close to a stationary point of the learning dynamics. In this purely theoretical contribution, we show that (for a very large class of generative models) the variational lower bound is at all stationary points of learning equal to a sum of entropies. For models with one set of latents and one set observed variables, the sum consists of three entropies: (A) the (average) entropy of the variational distributions, (B) the negative entropy of the model's prior distribution, and (C) the (expected) negative entropy of the observable distributions. The obtained result applies under realistic conditions including: finite numbers of data points, at any stationary points (including saddle points) and for any family of (well behaved) variational distributions. The class of generative models for which we show the equality to entropy sums contains many (and presumably most) standard generative models (including deep models). As concrete examples we discuss probabilistic PCA and Sigmoid Belief Networks. The prerequisites we use to show equality to entropy sums are relatively mild. Concretely, the distributions of a given generative model have to be of the exponential family (with constant base measure), and a model has to satisfy a parameterization criterion (which is usually fulfilled). Proving the equality of the ELBO to entropy sums at stationary points (under the stated conditions) is the main contribution of this work.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/28/2020

The Evidence Lower Bound of Variational Autoencoders Converges to a Sum of Three Entropies

The central objective function of a variational autoencoder (VAE) is its...
research
10/29/2019

A Prior of a Googol Gaussians: a Tensor Ring Induced Prior for Generative Models

Generative models produce realistic objects in many domains, including t...
research
06/01/2019

BreGMN: scaled-Bregman Generative Modeling Networks

The family of f-divergences is ubiquitously applied to generative modeli...
research
12/22/2020

Evolutionary Variational Optimization of Generative Models

We combine two popular optimization approaches to derive learning algori...
research
10/16/2018

Metropolis-Hastings view on variational inference and adversarial training

In this paper we propose to view the acceptance rate of the Metropolis-H...
research
06/12/2015

Bidirectional Helmholtz Machines

Efficient unsupervised training and inference in deep generative models ...
research
02/06/2023

Variational Information Pursuit for Interpretable Predictions

There is a growing interest in the machine learning community in develop...

Please sign up or login with your details

Forgot password? Click here to reset