Learning Shared Filter Bases for Efficient ConvNets

06/09/2020
by   Daeyeon Kim, et al.
0

Modern convolutional neural networks (ConvNets) achieve state-of-the-art performance for many computer vision tasks. However, such high performance requires millions of parameters and high computational costs. Recently, inspired by the iterative structure of modern ConvNets, such as ResNets, parameter sharing among repetitive convolution layers has been proposed to reduce the size of parameters. However, naive sharing of convolution filters poses many challenges such as overfitting and vanishing/exploding gradients. Furthermore, parameter sharing often increases computational complexity due to additional operations. In this paper, we propose to exploit the linear structure of convolution filters for effective and efficient sharing of parameters among iterative convolution layers. Instead of sharing convolution filters themselves, we hypothesize that a filter basis of linearly-decomposed convolution layers are more effective units for sharing parameters since a filter basis is an intrinsic and reusable building block constituting diverse high dimensional convolution filters. The representation power and peculiarity of individual convolution layers are further increased by adding a small number of layer-specific non-shared components to the filter basis. We show empirically that enforcing orthogonality to shared filter bases can mitigate the difficulty in training shared parameters. Experimental results show that our approach achieves significant reductions both in model parameters and computational costs while maintaining competitive, and often better, performance than non-shared baseline networks.

READ FULL TEXT
research
01/05/2018

Learning 3D-FilterMap for Deep Convolutional Neural Networks

We present a novel and compact architecture for deep Convolutional Neura...
research
02/20/2019

Spatially-Adaptive Filter Units for Compact and Efficient Deep Neural Networks

Convolutional neural networks excel in a number of computer vision tasks...
research
09/25/2020

Tied Block Convolution: Leaner and Better CNNs with Shared Thinner Filters

Convolution is the main building block of convolutional neural networks ...
research
05/17/2018

RotDCF: Decomposition of Convolutional Filters for Rotation-Equivariant Deep Networks

Explicit encoding of group actions in deep features makes it possible fo...
research
02/12/2018

DCFNet: Deep Neural Network with Decomposed Convolutional Filters

Filters in a Convolutional Neural Network (CNN) contain model parameters...
research
06/08/2017

Learning Local Receptive Fields and their Weight Sharing Scheme on Graphs

We propose a simple and generic layer formulation that extends the prope...
research
11/28/2017

WSNet: Compact and Efficient Networks with Weight Sampling

We present a new approach and a novel architecture, termed WSNet, for le...

Please sign up or login with your details

Forgot password? Click here to reset