Rescaling CNN through Learnable Repetition of Network Parameters

01/14/2021
by   Arnav Chavan, et al.
8

Deeper and wider CNNs are known to provide improved performance for deep learning tasks. However, most such networks have poor performance gain per parameter increase. In this paper, we investigate whether the gain observed in deeper models is purely due to the addition of more optimization parameters or whether the physical size of the network as well plays a role. Further, we present a novel rescaling strategy for CNNs based on learnable repetition of its parameters. Based on this strategy, we rescale CNNs without changing their parameter count, and show that learnable sharing of weights itself can provide significant boost in the performance of any given model without changing its parameter count. We show that small base networks when rescaled, can provide performance comparable to deeper networks with as low as 6 parameters of the deeper one. The relevance of weight sharing is further highlighted through the example of group-equivariant CNNs. We show that the significant improvements obtained with group-equivariant CNNs over the regular CNNs on classification problems are only partly due to the added equivariance property, and part of it comes from the learnable repetition of network weights. For rot-MNIST dataset, we show that up to 40 rotation equivariance could actually be due to just the learnt repetition of weights.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/25/2021

Exploiting Redundancy: Separable Group Convolutional Networks on Lie Groups

Group convolutional neural networks (G-CNNs) have been shown to increase...
research
11/23/2020

Learnable Gabor modulated complex-valued networks for orientation robustness

Robustness to transformation is desirable in many computer vision tasks,...
research
08/04/2019

Building Deep, Equivariant Capsule Networks

Capsule networks are constrained by their, relative, inability to deeper...
research
08/14/2016

About Pyramid Structure in Convolutional Neural Networks

Deep convolutional neural networks (CNN) brought revolution without any ...
research
09/05/2019

A Novel Design of Adaptive and Hierarchical Convolutional Neural Networks using Partial Reconfiguration on FPGA

Nowadays most research in visual recognition using Convolutional Neural ...
research
03/01/2023

Empowering Networks With Scale and Rotation Equivariance Using A Similarity Convolution

The translational equivariant nature of Convolutional Neural Networks (C...
research
06/14/2022

Prioritized Training on Points that are Learnable, Worth Learning, and Not Yet Learnt

Training on web-scale data can take months. But most computation and tim...

Please sign up or login with your details

Forgot password? Click here to reset