DO-Conv: Depthwise Over-parameterized Convolutional Layer

06/22/2020
by   Jinming Cao, et al.
0

Convolutional layers are the core building blocks of Convolutional Neural Networks (CNNs). In this paper, we propose to augment a convolutional layer with an additional depthwise convolution, where each input channel is convolved with a different 2D kernel. The composition of the two convolutions constitutes an over-parameterization, since it adds learnable parameters, while the resulting linear operation can be expressed by a single convolution layer. We refer to this depthwise over-parameterized convolutional layer as DO-Conv. We show with extensive experiments that the mere replacement of conventional convolutional layers with DO-Conv layers boosts the performance of CNNs on many classical vision tasks, such as image classification, detection, and segmentation. Moreover, in the inference phase, the depthwise convolution is folded into the conventional convolution, reducing the computation to be exactly equivalent to that of a convolutional layer without over-parameterization. As DO-Conv introduces performance gains without incurring any computational complexity increase for inference, we advocate it as an alternative to the conventional convolutional layer. We open-source a reference implementation of DO-Conv in Tensorflow, PyTorch and GluonCV at https://github.com/yangyanli/DO-Conv.

READ FULL TEXT
research
03/03/2021

An Alternative Practice of Tropical Convolution to Traditional Convolutional Neural Networks

Convolutional neural networks (CNNs) have been used in many machine lear...
research
07/26/2019

LinearConv: Regenerating Redundancy in Convolution Filters as Linear Combinations for Parameter Reduction

Convolutional Neural Networks (CNNs) show state-of-the-art performance i...
research
02/06/2020

Fixed smooth convolutional layer for avoiding checkerboard artifacts in CNNs

In this paper, we propose a fixed convolutional layer with an order of s...
research
10/12/2019

Context-Gated Convolution

As the basic building block of Convolutional Neural Networks (CNNs), the...
research
05/24/2019

Training decision trees as replacement for convolution layers

We present an alternative layer to convolution layers in convolutional n...
research
12/06/2021

DANets: Deep Abstract Networks for Tabular Data Classification and Regression

Tabular data are ubiquitous in real world applications. Although many co...
research
10/08/2021

Lightweight Convolutional Neural Networks By Hypercomplex Parameterization

Hypercomplex neural networks have proved to reduce the overall number of...

Please sign up or login with your details

Forgot password? Click here to reset