GSNs : Generative Stochastic Networks

03/18/2015
by   Guillaume Alain, et al.
0

We introduce a novel training principle for probabilistic models that is an alternative to maximum likelihood. The proposed Generative Stochastic Networks (GSN) framework is based on learning the transition operator of a Markov chain whose stationary distribution estimates the data distribution. Because the transition distribution is a conditional distribution generally involving a small move, it has fewer dominant modes, being unimodal in the limit of small moves. Thus, it is easier to learn, more like learning to perform supervised function approximation, with gradients that can be obtained by back-propagation. The theorems provided here generalize recent work on the probabilistic interpretation of denoising auto-encoders and provide an interesting justification for dependency networks and generalized pseudolikelihood (along with defining an appropriate joint distribution and sampling mechanism, even when the conditionals are not consistent). We study how GSNs can be used with missing inputs and can be used to sample subsets of variables given the rest. Successful experiments are conducted, validating these theoretical results, on two image datasets and with a particular architecture that mimics the Deep Boltzmann Machine Gibbs sampler but allows training to proceed with backprop, without the need for layerwise pretraining.

READ FULL TEXT

page 27

page 28

page 31

page 32

page 33

page 36

research
12/19/2013

Multimodal Transitions for Generative Stochastic Networks

Generative Stochastic Networks (GSNs) have been recently introduced as a...
research
03/20/2017

Learning to Generate Samples from Noise through Infusion Training

In this work, we investigate a novel training procedure to learn a gener...
research
12/22/2014

Generative Class-conditional Autoencoders

Recent work by Bengio et al. (2013) proposes a sampling procedure for de...
research
07/02/2021

Reconsidering Dependency Networks from an Information Geometry Perspective

Dependency networks (Heckerman et al., 2000) are potential probabilistic...
research
03/20/2013

Integrating Probabilistic Rules into Neural Networks: A Stochastic EM Learning Algorithm

The EM-algorithm is a general procedure to get maximum likelihood estima...
research
03/09/2023

Learning Stationary Markov Processes with Contrastive Adjustment

We introduce a new optimization algorithm, termed contrastive adjustment...
research
11/07/2017

Variational Walkback: Learning a Transition Operator as a Stochastic Recurrent Net

We propose a novel method to directly learn a stochastic transition oper...

Please sign up or login with your details

Forgot password? Click here to reset