PredProp: Bidirectional Stochastic Optimization with Precision Weighted Predictive Coding

11/16/2021
by   André Ofner, et al.
0

We present PredProp, a method for bidirectional, parallel and local optimisation of weights, activities and precision in neural networks. PredProp jointly addresses inference and learning, scales learning rates dynamically and weights gradients by the curvature of the loss function by optimizing prediction error precision. PredProp optimizes network parameters with Stochastic Gradient Descent and error forward propagation based strictly on prediction errors and variables locally available to each layer. Neighboring layers optimise shared activity variables so that prediction errors can propagate forward in the network, while predictions propagate backwards. This process minimises the negative Free Energy, or evidence lower bound of the entire network. We show that networks trained with PredProp resemble gradient based predictive coding when the number of weights between neighboring activity variables is one. In contrast to related work, PredProp generalizes towards backward connections of arbitrary depth and optimizes precision for any deep network architecture. Due to the analogy between prediction error precision and the Fisher information for each layer, PredProp implements a form of Natural Gradient Descent. When optimizing DNN models, layer-wise PredProp renders the model a bidirectional predictive coding network. Alternatively DNNs can parameterize the weights between two activity variables. We evaluate PredProp for dense DNNs on simple inference, learning and combined tasks. We show that, without an explicit sampling step in the network, PredProp implements a form of variational inference that allows to learn disentangled embeddings from low amounts of data and leave evaluation on more complex tasks and datasets to future work.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/12/2021

Predictive coding, precision and natural gradients

There is an increasing convergence between biologically plausible comput...
research
05/29/2020

Bayesian Neural Network via Stochastic Gradient Descent

The goal of bayesian approach used in variational inference is to minimi...
research
07/03/2018

Stochastic Layer-Wise Precision in Deep Neural Networks

Low precision weights, activations, and gradients have been proposed as ...
research
12/02/2021

Differentiable Generalised Predictive Coding

This paper deals with differentiable dynamical models congruent with neu...
research
10/02/2020

Relaxing the Constraints on Predictive Coding Models

Predictive coding is an influential theory of cortical function which po...
research
05/29/2023

Deep Predictive Coding with Bi-directional Propagation for Classification and Reconstruction

This paper presents a new learning algorithm, termed Deep Bi-directional...
research
08/14/2018

Generalization of Equilibrium Propagation to Vector Field Dynamics

The biological plausibility of the backpropagation algorithm has long be...

Please sign up or login with your details

Forgot password? Click here to reset