Network size and weights size for memorization with two-layers neural networks

06/04/2020
by   Sébastien Bubeck, et al.
12

In 1988, Eric B. Baum showed that two-layers neural networks with threshold activation function can perfectly memorize the binary labels of n points in general position in R^d using only n/d neurons. We observe that with ReLU networks, using four times as many neurons one can fit arbitrary real labels. Moreover, for approximate memorization up to error ϵ, the neural tangent kernel can also memorize with only O(n/d·log(1/ϵ) ) neurons (assuming that the data is well dispersed too). We show however that these constructions give rise to networks where the magnitude of the neurons' weights are far from optimal. In contrast we propose a new training procedure for ReLU networks, based on complex (as opposed to real) recombination of the neurons, for which we show approximate memorization with both O(n/d·log(1/ϵ)/ϵ) neurons, as well as nearly-optimal size of the weights.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/27/2022

Expressive power of binary and ternary neural networks

We show that deep sparse ReLU networks with ternary weights and deep ReL...
research
06/14/2021

An Exponential Improvement on the Memorization Capacity of Deep Threshold Networks

It is well known that modern deep neural networks are powerful enough to...
research
06/18/2022

Coin Flipping Neural Networks

We show that neural networks with access to randomness can outperform de...
research
11/03/2021

Regularization by Misclassification in ReLU Neural Networks

We study the implicit bias of ReLU neural networks trained by a variant ...
research
04/04/2022

Training Fully Connected Neural Networks is ∃ℝ-Complete

We consider the algorithmic problem of finding the optimal weights and b...
research
03/17/2019

Evolving and Understanding Sparse Deep Neural Networks using Cosine Similarity

Training sparse neural networks with adaptive connectivity is an active ...
research
08/06/2019

Refining the Structure of Neural Networks Using Matrix Conditioning

Deep learning models have proven to be exceptionally useful in performin...

Please sign up or login with your details

Forgot password? Click here to reset