Understanding Deep Architectures using a Recursive Convolutional Network

12/06/2013
by   David Eigen, et al.
0

A key challenge in designing convolutional network models is sizing them appropriately. Many factors are involved in these decisions, including number of layers, feature maps, kernel sizes, etc. Complicating this further is the fact that each of these influence not only the numbers and dimensions of the activation units, but also the total number of parameters. In this paper we focus on assessing the independent contributions of three of these linked variables: The numbers of layers, feature maps, and parameters. To accomplish this, we employ a recursive convolutional network whose weights are tied between layers; this allows us to vary each of the three factors in a controlled setting. We find that while increasing the numbers of layers and parameters each have clear benefit, the number of feature maps (and hence dimensionality of the representation) appears ancillary, and finds most of its benefit through the introduction of more weights. Our results (i) empirically confirm the notion that adding layers alone increases computational power, within the context of convolutional layers, and (ii) suggest that precise sizing of convolutional feature map dimensions is itself of little concern; more attention should be paid to the number of parameters in these layers instead.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/29/2019

Size-free generalization bounds for convolutional neural networks

We prove bounds on the generalization error of convolutional networks. T...
research
04/17/2021

Towards Efficient Convolutional Network Models with Filter Distribution Templates

Increasing number of filters in deeper layers when feature maps are decr...
research
07/20/2020

ThriftyNets : Convolutional Neural Networks with Tiny Parameter Budget

Typical deep convolutional architectures present an increasing number of...
research
05/27/2019

SpecNet: Spectral Domain Convolutional Neural Network

The memory consumption of most Convolutional Neural Network (CNN) archit...
research
11/22/2018

TGANv2: Efficient Training of Large Models for Video Generation with Multiple Subsampling Layers

In this paper, we propose a novel method to efficiently train a Generati...
research
08/18/2021

Towards Deep and Efficient: A Deep Siamese Self-Attention Fully Efficient Convolutional Network for Change Detection in VHR Images

Recently, FCNs have attracted widespread attention in the CD field. In p...

Please sign up or login with your details

Forgot password? Click here to reset