Provable Memorization via Deep Neural Networks using Sub-linear Parameters

10/26/2020
by   Sejun Park, et al.
0

It is known that Θ(N) parameters are sufficient for neural networks to memorize arbitrary N input-label pairs. By exploiting depth, we show that Θ(N^2/3) parameters suffice to memorize N pairs, under a mild condition on the separation of input points. In particular, deeper networks (even with width 3) are shown to memorize more pairs than shallow networks, which also agrees with the recent line of works on the benefits of depth for function approximation. We also provide empirical results that support our theoretical findings.

READ FULL TEXT

page 10

page 13

research
03/15/2021

Function approximation by deep neural networks with parameters {0,±1/2, ± 1, 2}

In this paper it is shown that C_β-smooth functions can be approximated ...
research
01/31/2021

The Connection Between Approximation, Depth Separation and Learnability in Neural Networks

Several recent works have shown separation results between deep neural n...
research
02/08/2022

Width is Less Important than Depth in ReLU Neural Networks

We solve an open question from Lu et al. (2017), by showing that any tar...
research
12/04/2021

Optimization-Based Separations for Neural Networks

Depth separation results propose a possible theoretical explanation for ...
research
10/31/2016

Depth-Width Tradeoffs in Approximating Natural Functions with Neural Networks

We provide several new depth-based separation results for feed-forward n...
research
10/27/2022

On the Approximation and Complexity of Deep Neural Networks to Invariant Functions

Recent years have witnessed a hot wave of deep neural networks in variou...
research
05/11/2023

Rethink Depth Separation with Intra-layer Links

The depth separation theory is nowadays widely accepted as an effective ...

Please sign up or login with your details

Forgot password? Click here to reset