Early Inference in Energy-Based Models Approximates Back-Propagation

10/09/2015
by   Yoshua Bengio, et al.
0

We show that Langevin MCMC inference in an energy-based model with latent variables has the property that the early steps of inference, starting from a stationary point, correspond to propagating error gradients into internal layers, similarly to back-propagation. The error that is back-propagated is with respect to visible units that have received an outside driving force pushing them away from the stationary point. Back-propagated error gradients correspond to temporal derivatives of the activation of hidden units. This observation could be an element of a theory for explaining how brains perform credit assignment in deep hierarchies as efficiently as back-propagation does. In this theory, the continuous-valued latent variables correspond to averaged voltage potential (across time, spikes, and possibly neurons in the same minicolumn), and neural computation corresponds to approximate inference and error back-propagation at the same time.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/16/2016

Equilibrium Propagation: Bridging the Gap Between Energy-Based Models and Backpropagation

We introduce Equilibrium Propagation, a learning framework for energy-ba...
research
12/23/2014

Difference Target Propagation

Back-propagation has been the workhorse of recent successes of deep lear...
research
06/25/2021

Conjugate Energy-Based Models

In this paper, we propose conjugate energy-based models (CEBMs), a new c...
research
02/13/2018

Bayesian model assessment: Use of conditional vs marginal likelihoods

Typical Bayesian methods for models with latent variables (or random eff...
research
06/12/2020

Learning Energy-based Model with Flow-based Backbone by Neural Transport MCMC

Learning energy-based model (EBM) requires MCMC sampling of the learned ...
research
05/16/2019

Formal derivation of Mesh Neural Networks with their Forward-Only gradient Propagation

This paper proposes the Mesh Neural Network (MNN), a novel architecture ...
research
02/03/2015

Deep Joint Task Learning for Generic Object Extraction

This paper investigates how to extract objects-of-interest without relyi...

Please sign up or login with your details

Forgot password? Click here to reset