Rediscovering Deep Neural Networks in Finite-State Distributions

09/26/2018
by   Amir Emad Marvasti, et al.
4

We propose a new way of thinking about deep neural networks, in which the linear and non-linear components of the network are naturally derived and justified in terms of principles in probability theory. In particular, the models constructed in our framework assign probabilities to uncertain realizations, leading to Kullback-Leibler Divergence (KLD) as the linear layer. In our model construction, we also arrive at a structure similar to ReLU activation supported with Bayes' theorem. The non-linearities in our framework are normalization layers with ReLU and Sigmoid as element-wise approximations. Additionally, the pooling function is derived as a marginalization of spatial random variables according to the mechanics of the framework. As such, Max Pooling is an approximation to the aforementioned marginalization process. Since our models are comprised of finite state distributions (FSD) as variables and parameters, exact computation of information-theoretic quantities such as entropy and KLD is possible, thereby providing more objective measures to analyze networks. Unlike existing designs that rely on heuristics, the proposed framework restricts subjective interpretations of CNNs and sheds light on the functionality of neural networks from a completely new perspective.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/01/2021

Approximation Properties of Deep ReLU CNNs

This paper is devoted to establishing L^2 approximation properties for d...
research
04/09/2019

Approximation in L^p(μ) with deep ReLU neural networks

We discuss the expressive power of neural networks which use the non-smo...
research
09/13/2019

Shapley Interpretation and Activation in Neural Networks

We propose a novel Shapley value approach to help address neural network...
research
03/02/2022

The Theoretical Expressiveness of Maxpooling

Over the decade since deep neural networks became state of the art image...
research
06/08/2023

Normalization-Equivariant Neural Networks with Application to Image Denoising

In many information processing systems, it may be desirable to ensure th...
research
11/03/2021

A Johnson–Lindenstrauss Framework for Randomly Initialized CNNs

How does the geometric representation of a dataset change after the appl...
research
09/18/2017

A Probabilistic Framework for Nonlinearities in Stochastic Neural Networks

We present a probabilistic framework for nonlinearities, based on doubly...

Please sign up or login with your details

Forgot password? Click here to reset