Feed-forward Uncertainty Propagation in Belief and Neural Networks

03/28/2018
by   Alexander Shekhovtsov, et al.
0

We propose a feed-forward inference method applicable to belief and neural networks. In a belief network, the method estimates an approximate factorized posterior of all hidden units given the input. In neural networks the method propagates uncertainty of the input through all the layers. In neural networks with injected noise, the method analytically takes into account uncertainties resulting from this noise. Such feed-forward analytic propagation is differentiable in parameters and can be trained end-to-end. Compared to standard NN, which can be viewed as propagating only the means, we propagate the mean and variance. The method can be useful in all scenarios that require knowledge of the neuron statistics, e.g. when dealing with uncertain inputs, considering sigmoid activations as probabilities of Bernoulli units, training the models regularized by injected noise (dropout) or estimating activation statistics over the dataset (as needed for normalization methods). In the experiments we show the possible utility of the method in all these tasks as well as its current limitations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/28/2018

Normalization of Neural Networks using Analytic Variance Propagation

We address the problem of estimating statistics of hidden units in a neu...
research
06/11/2018

When and where do feed-forward neural networks learn localist representations?

According to parallel distributed processing (PDP) theory in psychology,...
research
05/19/2022

Causal Discovery and Injection for Feed-Forward Neural Networks

Neural networks have proven to be effective at solving a wide range of p...
research
05/22/2018

Expectation propagation: a probabilistic view of Deep Feed Forward Networks

We present a statistical mechanics model of deep feed forward neural net...
research
03/30/2023

Optimal Input Gain: All You Need to Supercharge a Feed-Forward Neural Network

Linear transformation of the inputs alters the training performance of f...
research
09/17/2022

Introspective Learning : A Two-Stage Approach for Inference in Neural Networks

In this paper, we advocate for two stages in a neural network's decision...
research
10/19/2011

Readouts for Echo-state Networks Built using Locally Regularized Orthogonal Forward Regression

Echo state network (ESN) is viewed as a temporal non-orthogonal expansio...

Please sign up or login with your details

Forgot password? Click here to reset