Capacity allocation through neural network layers

02/22/2019
by   Jonathan Donier, et al.
0

Capacity analysis has been recently introduced as a way to analyze how linear models distribute their modelling capacity across the input space. In this paper, we extend the notion of capacity allocation to the case of neural networks with non-linear layers. We show that under some hypotheses the problem is equivalent to linear capacity allocation, within some extended input space that factors in the non-linearities. We introduce the notion of layer decoupling, which quantifies the degree to which a non-linear activation decouples its outputs, and show that it plays a central role in capacity allocation through layers. In the highly non-linear limit where decoupling is total, we show that the propagation of capacity throughout the layers follows a simple markovian rule, which turns into a diffusion PDE in the limit of deep networks with residual layers. This allows us to recover some known results about deep neural networks, such as the size of the effective receptive field, or why ResNets avoid the shattering problem.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/11/2019

Scaling up deep neural networks: a capacity allocation perspective

Following the recent work on capacity allocation, we formulate the conje...
research
02/12/2019

Capacity allocation analysis of neural networks: A tool for principled architecture design

Designing neural network architectures is a task that lies somewhere bet...
research
02/04/2020

A Deep Conditioning Treatment of Neural Networks

We study the role of depth in training randomly initialized overparamete...
research
10/17/2017

Spontaneous Symmetry Breaking in Neural Networks

We propose a framework to understand the unprecedented performance and r...
research
07/15/2020

From deep to Shallow: Equivalent Forms of Deep Networks in Reproducing Kernel Krein Space and Indefinite Support Vector Machines

In this paper we explore a connection between deep networks and learning...
research
03/18/2021

The Low-Rank Simplicity Bias in Deep Networks

Modern deep neural networks are highly over-parameterized compared to th...
research
01/02/2019

The capacity of feedforward neural networks

A long standing open problem in the theory of neural networks is the dev...

Please sign up or login with your details

Forgot password? Click here to reset