Reweighted Expectation Maximization

06/13/2019
by   Adji B. Dieng, et al.
0

Training deep generative models with maximum likelihood remains a challenge. The typical workaround is to use variational inference (VI) and maximize a lower bound to the log marginal likelihood of the data. Variational auto-encoders (VAEs) adopt this approach. They further amortize the cost of inference by using a recognition network to parameterize the variational family. Amortized VI scales approximate posterior inference in deep generative models to large datasets. However it introduces an amortization gap and leads to approximate posteriors of reduced expressivity due to the problem known as posterior collapse. In this paper, we consider expectation maximization (EM) as a paradigm for fitting deep generative models. Unlike VI, EM directly maximizes the log marginal likelihood of the data. We rediscover the importance weighted auto-encoder (IWAE) as an instance of EM and propose a new EM-based algorithm for fitting deep generative models called reweighted expectation maximization (REM). REM learns better generative models than the IWAE by decoupling the learning dynamics of the generative model and the recognition network using a separate expressive proposal found by moment matching. We compared REM to the VAE and the IWAE on several density estimation benchmarks and found it leads to significantly better performance as measured by log-likelihood.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/17/2020

Analytical Probability Distributions and EM-Learning for Deep Generative Networks

Deep Generative Networks (DGNs) with probabilistic modeling of their out...
research
12/07/2012

Layer-wise learning of deep generative models

When using deep, multi-layered architectures to build generative models ...
research
06/12/2015

Bidirectional Helmholtz Machines

Efficient unsupervised training and inference in deep generative models ...
research
07/11/2018

VFunc: a Deep Generative Model for Functions

We introduce a deep generative model for functions. Our model provides a...
research
04/22/2020

Provably robust deep generative models

Recent work in adversarial attacks has developed provably robust methods...
research
03/20/2016

Joint Stochastic Approximation learning of Helmholtz Machines

Though with progress, model learning and performing posterior inference ...
research
06/10/2019

Multi-objects Generation with Amortized Structural Regularization

Deep generative models (DGMs) have shown promise in image generation. Ho...

Please sign up or login with your details

Forgot password? Click here to reset