Multigrid-in-Channels Architectures for Wide Convolutional Neural Networks

06/11/2020
by   Jonathan Ephrath, et al.
0

We present a multigrid approach that combats the quadratic growth of the number of parameters with respect to the number of channels in standard convolutional neural networks (CNNs). It has been shown that there is a redundancy in standard CNNs, as networks with much sparser convolution operators can yield similar performance to full networks. The sparsity patterns that lead to such behavior, however, are typically random, hampering hardware efficiency. In this work, we present a multigrid-in-channels approach for building CNN architectures that achieves full coupling of the channels, and whose number of parameters is linearly proportional to the width of the network. To this end, we replace each convolution layer in a generic CNN with a multilevel layer consisting of structured (i.e., grouped) convolutions. Our examples from supervised image classification show that applying this strategy to residual networks and MobileNetV2 considerably reduces the number of parameters without negatively affecting accuracy. Therefore, we can widen networks without dramatically increasing the number of parameters or operations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/17/2020

Multigrid-in-Channels Neural Network Architectures

We present a multigrid-in-channels (MGIC) approach that tackles the quad...
research
05/11/2019

Training CNNs with Selective Allocation of Channels

Recent progress in deep convolutional neural networks (CNNs) have enable...
research
11/07/2020

Depthwise Multiception Convolution for Reducing Network Parameters without Sacrificing Accuracy

Deep convolutional neural networks have been proven successful in multip...
research
10/15/2017

CNNComparator: Comparative Analytics of Convolutional Neural Networks

Convolutional neural networks (CNNs) are widely used in many image recog...
research
06/08/2016

Convolutional Neural Fabrics

Despite the success of CNNs, selecting the optimal architecture for a gi...
research
02/10/2023

DNArch: Learning Convolutional Neural Architectures by Backpropagation

We present Differentiable Neural Architectures (DNArch), a method that j...
research
09/04/2020

ACDC: Weight Sharing in Atom-Coefficient Decomposed Convolution

Convolutional Neural Networks (CNNs) are known to be significantly over-...

Please sign up or login with your details

Forgot password? Click here to reset