Variational Walkback: Learning a Transition Operator as a Stochastic Recurrent Net

11/07/2017
by   Anirudh Goyal, et al.
0

We propose a novel method to directly learn a stochastic transition operator whose repeated application provides generated samples. Traditional undirected graphical models approach this problem indirectly by learning a Markov chain model whose stationary distribution obeys detailed balance with respect to a parameterized energy function. The energy function is then modified so the model and data distributions match, with no guarantee on the number of steps required for the Markov chain to converge. Moreover, the detailed balance condition is highly restrictive: energy based models corresponding to neural networks must have symmetric weights, unlike biological neural circuits. In contrast, we develop a method for directly learning arbitrarily parameterized transition operators capable of expressing non-equilibrium stationary distributions that violate detailed balance, thereby enabling us to learn more biologically plausible asymmetric neural networks and more general non-energy based dynamical systems. The proposed training objective, which we derive via principled variational methods, encourages the transition operator to "walk back" in multi-step trajectories that start at data-points, as quickly as possible back to the original data points. We present a series of experimental results illustrating the soundness of the proposed approach, Variational Walkback (VW), on the MNIST, CIFAR-10, SVHN and CelebA datasets, demonstrating superior samples compared to earlier attempts to learn a transition operator. We also show that although each rapid training trajectory is limited to a finite but variable number of steps, our transition operator continues to generate good samples well past the length of such trajectories, thereby demonstrating the match of its non-equilibrium stationary distribution to the data distribution. Source Code: http://github.com/anirudh9119/walkback_nips17

READ FULL TEXT

page 13

page 19

page 20

page 21

page 22

page 23

page 24

page 25

research
12/19/2013

Multimodal Transitions for Generative Stochastic Networks

Generative Stochastic Networks (GSNs) have been recently introduced as a...
research
03/20/2017

Learning to Generate Samples from Noise through Infusion Training

In this work, we investigate a novel training procedure to learn a gener...
research
12/22/2014

Generative Class-conditional Autoencoders

Recent work by Bengio et al. (2013) proposes a sampling procedure for de...
research
06/26/2020

On the convergence of the Metropolis algorithm with fixed-order updates for multivariate binary probability distributions

The Metropolis algorithm is arguably the most fundamental Markov chain M...
research
02/27/2019

An evolutionary model that satisfies detailed balance

We propose a class of evolutionary models that involves an arbitrary exc...
research
03/09/2015

Mathematical understanding of detailed balance condition violation and its application to Langevin dynamics

We develop an efficient sampling method by simulating Langevin dynamics ...
research
03/18/2015

GSNs : Generative Stochastic Networks

We introduce a novel training principle for probabilistic models that is...

Please sign up or login with your details

Forgot password? Click here to reset