Imposing Gaussian Pre-Activations in a Neural Network

05/24/2022
by   Pierre Wolinski, et al.
0

The goal of the present work is to propose a way to modify both the initialization distribution of the weights of a neural network and its activation function, such that all pre-activations are Gaussian. We propose a family of pairs initialization/activation, where the activation functions span a continuum from bounded functions (such as Heaviside or tanh) to the identity function. This work is motivated by the contradiction between existing works dealing with Gaussian pre-activations: on one side, the works in the line of the Neural Tangent Kernels and the Edge of Chaos are assuming it, while on the other side, theoretical and experimental results challenge this hypothesis. The family of pairs initialization/activation we are proposing will help us to answer this hot question: is it desirable to have Gaussian pre-activations in a neural network?

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/18/2020

A Survey on Activation Functions and their relation with Xavier and He Normal Initialization

In artificial neural network, the activation function and the weight ini...
research
02/22/2021

Elementary superexpressive activations

We call a finite family of activation functions superexpressive if any m...
research
08/16/2019

Effect of Activation Functions on the Training of Overparametrized Neural Nets

It is well-known that overparametrized neural networks trained using gra...
research
12/17/2020

Guiding Neural Network Initialization via Marginal Likelihood Maximization

We propose a simple, data-driven approach to help guide hyperparameter s...
research
10/03/2017

Training Feedforward Neural Networks with Standard Logistic Activations is Feasible

Training feedforward neural networks with standard logistic activations ...
research
10/19/2020

Stationary Activations for Uncertainty Calibration in Deep Learning

We introduce a new family of non-linear neural network activation functi...
research
11/30/2021

Beyond Periodicity: Towards a Unifying Framework for Activations in Coordinate-MLPs

Coordinate-MLPs are emerging as an effective tool for modeling multidime...

Please sign up or login with your details

Forgot password? Click here to reset