Reweighted Wake-Sleep

06/11/2014
by   Jörg Bornschein, et al.
0

Training deep directed graphical models with many hidden variables and performing inference remains a major challenge. Helmholtz machines and deep belief networks are such models, and the wake-sleep algorithm has been proposed to train them. The wake-sleep algorithm relies on training not just the directed generative model but also a conditional generative model (the inference network) that runs backward from visible to latent, estimating the posterior distribution of latent given visible. We propose a novel interpretation of the wake-sleep algorithm which suggests that better estimators of the gradient can be obtained by sampling latent variables multiple times from the inference network. This view is based on importance sampling as an estimator of the likelihood, with the approximate inference network as a proposal distribution. This interpretation is confirmed experimentally, showing that better likelihood can be achieved with this reweighted wake-sleep procedure. Based on this interpretation, we propose that a sigmoidal belief network is not sufficiently powerful for the layers of the inference network in order to recover a good estimator of the posterior distribution of latent variables. Our experiments show that using a more powerful layer model, such as NADE, yields substantially better generative models.

READ FULL TEXT

page 7

page 8

research
12/04/2019

Learning Deep Generative Models with Short Run Inference Dynamics

This paper studies the fundamental problem of learning deep generative m...
research
05/18/2023

Massively Parallel Reweighted Wake-Sleep

Reweighted wake-sleep (RWS) is a machine learning method for performing ...
research
05/25/2023

Learning Directed Graphical Models with Optimal Transport

Estimating the parameters of a probabilistic directed graphical model fr...
research
09/22/2015

Learning Wake-Sleep Recurrent Attention Models

Despite their success, convolutional neural networks are computationally...
research
07/06/2020

Learning to learn generative programs with Memoised Wake-Sleep

We study a class of neuro-symbolic generative models in which neural net...
research
06/12/2018

Explaining and Generalizing Back-Translation through Wake-Sleep

Back-translation has become a commonly employed heuristic for semi-super...
research
08/15/2020

Natural Wake-Sleep Algorithm

The benefits of using the natural gradient are well known in a wide rang...

Please sign up or login with your details

Forgot password? Click here to reset