Removable and/or Repeated Units Emerge in Overparametrized Deep Neural Networks

12/10/2019
by   Stephen Casper, et al.
0

Deep neural networks (DNNs) perform well on a variety of tasks despite the fact that most networks used in practice are vastly overparametrized and even capable of perfectly fitting randomly labeled data. Recent evidence suggests that developing compressible representations is key for adjusting the complexity of overparametrized networks to the task at hand. In this paper, we provide new empirical evidence that supports this hypothesis by identifying two types of units that emerge when the network's width is increased: removable units which can be dropped out of the network without significant change to the output and repeated units whose activities are highly correlated with other units. The emergence of these units implies capacity constraints as the function the network represents could be expressed by a smaller network without these units. In a series of experiments with AlexNet, ResNet and Inception networks in the CIFAR-10 and ImageNet datasets, and also using shallow networks with synthetic data, we show that DNNs consistently increase either the number of removable units, repeated units, or both at greater widths for a comprehensive set of hyperparameters. These results suggest that the mechanisms by which networks in the deep learning regime adjust their complexity operate at the unit level and highlight the need for additional research into what drives the emergence of such units.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/27/2018

Understanding the Importance of Single Directions via Representative Substitution

Understanding the internal representations of deep neural networks (DNNs...
research
07/27/2018

Diverse feature visualizations reveal invariances in early layers of deep neural networks

Visualizing features in deep neural networks (DNNs) can help understandi...
research
04/27/2020

Interpretation of Deep Temporal Representations by Selective Visualization of Internally Activated Units

Recently deep neural networks demonstrate competitive performances in cl...
research
03/03/2020

Selectivity considered harmful: evaluating the causal impact of class selectivity in DNNs

Class selectivity, typically defined as how different a neuron's respons...
research
03/14/2018

Building Sparse Deep Feedforward Networks using Tree Receptive Fields

Sparse connectivity is an important factor behind the success of convolu...
research
03/07/2017

Sharing Residual Units Through Collective Tensor Factorization in Deep Neural Networks

Residual units are wildly used for alleviating optimization difficulties...

Please sign up or login with your details

Forgot password? Click here to reset