On Measuring Excess Capacity in Neural Networks

02/16/2022
by   Florian Graf, et al.
0

We study the excess capacity of deep networks in the context of supervised classification. That is, given a capacity measure of the underlying hypothesis class – in our case, Rademacher complexity – how much can we (a-priori) constrain this class while maintaining an empirical error comparable to the unconstrained setting. To assess excess capacity in modern architectures, we first extend an existing generalization bound to accommodate function composition and addition, as well as the specific structure of convolutions. This then facilitates studying residual networks through the lens of the accompanying capacity measure. The key quantities driving this measure are the Lipschitz constants of the layers and the (2,1) group norm distance to the initializations of the convolution weights. We show that these quantities (1) can be kept surprisingly small and, (2) since excess capacity unexpectedly increases with task difficulty, this points towards an unnecessarily large capacity of unconstrained models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/22/2019

Global Capacity Measures for Deep ReLU Networks via Path Sampling

Classical results on the statistical complexity of linear models have co...
research
09/19/2018

Capacity Control of ReLU Neural Networks by Basis-path Norm

Recently, path norm was proposed as a new capacity measure for neural ne...
research
11/05/2017

Fisher-Rao Metric, Geometry, and Complexity of Neural Networks

We study the relationship between geometry and capacity measures for dee...
research
11/11/2019

An empirical study of the relation between network architecture and complexity

In this preregistration submission, we propose an empirical study of how...
research
10/21/2020

High-Capacity Complex Convolutional Neural Networks For I/Q Modulation Classification

I/Q modulation classification is a unique pattern recognition problem as...
research
01/02/2019

The capacity of feedforward neural networks

A long standing open problem in the theory of neural networks is the dev...
research
01/16/2013

Big Neural Networks Waste Capacity

This article exposes the failure of some big neural networks to leverage...

Please sign up or login with your details

Forgot password? Click here to reset