Drastically Reducing the Number of Trainable Parameters in Deep CNNs by Inter-layer Kernel-sharing

10/23/2022
by   Alireza Azadbakht, et al.
0

Deep convolutional neural networks (DCNNs) have become the state-of-the-art (SOTA) approach for many computer vision tasks: image classification, object detection, semantic segmentation, etc. However, most SOTA networks are too large for edge computing. Here, we suggest a simple way to reduce the number of trainable parameters and thus the memory footprint: sharing kernels between multiple convolutional layers. Kernel-sharing is only possible between “isomorphic" layers, i.e.layers having the same kernel size, input and output channels. This is typically the case inside each stage of a DCNN. Our experiments on CIFAR-10 and CIFAR-100, using the ConvMixer and SE-ResNet architectures show that the number of parameters of these models can drastically be reduced with minimal cost on accuracy. The resulting networks are appealing for certain edge computing applications that are subject to severe memory constraints, and even more interesting if leveraging "frozen weights" hardware accelerators. Kernel-sharing is also an efficient regularization method, which can reduce overfitting. The codes are publicly available at https://github.com/AlirezaAzadbakht/kernel-sharing.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/22/2019

Deep Anchored Convolutional Neural Networks

Convolutional Neural Networks (CNNs) have been proven to be extremely su...
research
11/10/2022

MGiaD: Multigrid in all dimensions. Efficiency and robustness by coarsening in resolution and channel dimensions

Current state-of-the-art deep neural networks for image classification a...
research
12/01/2022

ResNet Structure Simplification with the Convolutional Kernel Redundancy Measure

Deep learning, especially convolutional neural networks, has triggered a...
research
06/08/2016

Convolutional Neural Fabrics

Despite the success of CNNs, selecting the optimal architecture for a gi...
research
04/11/2023

A priori compression of convolutional neural networks for wave simulators

Convolutional neural networks are now seeing widespread use in a variety...
research
06/21/2023

Efficient ResNets: Residual Network Design

ResNets (or Residual Networks) are one of the most commonly used models ...
research
11/24/2019

Pixel Adaptive Filtering Units

State-of-the-art methods for computer vision rely heavily on the transla...

Please sign up or login with your details

Forgot password? Click here to reset