A Neural Network Based on First Principles

02/18/2020
by   Paul M. Baggenstoss, et al.
0

In this paper, a Neural network is derived from first principles, assuming only that each layer begins with a linear dimension-reducing transformation. The approach appeals to the principle of Maximum Entropy (MaxEnt) to find the posterior distribution of the input data of each layer, conditioned on the layer output variables. This posterior has a well-defined mean, the conditional mean estimator, that is calculated using a type of neural network with theoretically-derived activation functions similar to sigmoid, softplus, and relu. This implicitly provides a theoretical justification for their use. A theorem that finds the conditional distribution and conditional mean estimator under the MaxEnt prior is proposed, unifying results for special cases. Combining layers results in an auto-encoder with conventional feed-forward analysis network and a type of linear Bayesian belief network in the reconstruction path.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/14/2023

Improved Auto-Encoding using Deterministic Projected Belief Networks

In this paper, we exploit the unique properties of a deterministic proje...
research
04/13/2021

Maximum Entropy Auto-Encoding

In this paper, it is shown that an auto-encoder using optimal reconstruc...
research
10/28/2020

Estimating Multiplicative Relations in Neural Networks

Universal approximation theorem suggests that a shallow neural network c...
research
04/25/2022

Using the Projected Belief Network at High Dimensions

The projected belief network (PBN) is a layered generative network (LGN)...
research
06/08/2023

Posterior Collapse in Linear Conditional and Hierarchical Variational Autoencoders

The posterior collapse phenomenon in variational autoencoders (VAEs), wh...
research
04/25/2020

Compromise-free Bayesian neural networks

We conduct a thorough analysis of the relationship between the out-of-sa...
research
06/30/2020

Deriving Neural Network Design and Learning from the Probabilistic Framework of Chain Graphs

The last decade has witnessed a boom of neural network (NN) research and...

Please sign up or login with your details

Forgot password? Click here to reset