Neural networks are a priori biased towards Boolean functions with low entropy

09/25/2019
by   Chris Mingard, et al.
0

Understanding the inductive bias of neural networks is critical to explaining their ability to generalise. Here, for one of the simplest neural networks -- a single-layer perceptron with n input neurons, one output neuron, and no threshold bias term -- we prove that upon random initialisation of weights, the a priori probability P(t) that it represents a Boolean function that classifies t points in {0,1}^n as 1 has a remarkably simple form: P(t) = 2^-n for 0≤ t < 2^n. Since a perceptron can express far fewer Boolean functions with small or large values of t (low "entropy") than with intermediate values of t (high "entropy") there is, on average, a strong intrinsic a-priori bias towards individual functions with low entropy. Furthermore, within a class of functions with fixed t, we often observe a further intrinsic bias towards functions of lower complexity. Finally, we prove that, regardless of the distribution of inputs, the bias towards low entropy becomes monotonically stronger upon adding ReLU layers, and empirically show that increasing the variance of the bias term has a similar effect.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/22/2018

Deep learning generalizes because the parameter-function map is biased towards simple functions

Deep neural networks generalize remarkably well without explicit regular...
research
03/27/2019

Fourier Entropy-Influence Conjecture for Random Linear Threshold Functions

The Fourier-Entropy Influence (FEI) Conjecture states that for any Boole...
research
05/28/2019

Average Bias and Polynomial Sources

We identify a new notion of pseudorandomness for randomness sources, whi...
research
06/10/2022

Intrinsic dimensionality and generalization properties of the ℛ-norm inductive bias

We study the structural and statistical properties of ℛ-norm minimizing ...
research
06/10/2018

On the Fourier Entropy Influence Conjecture for Extremal Classes

The Fourier Entropy-Influence (FEI) Conjecture of Friedgut and Kalai sta...
research
11/30/2021

Neuron with Steady Response Leads to Better Generalization

Regularization can mitigate the generalization gap between training and ...
research
05/02/2019

Full-Jacobian Representation of Neural Networks

Non-linear functions such as neural networks can be locally approximated...

Please sign up or login with your details

Forgot password? Click here to reset