MGiaD: Multigrid in all dimensions. Efficiency and robustness by coarsening in resolution and channel dimensions

11/10/2022
by   Antonia van Betteray, et al.
0

Current state-of-the-art deep neural networks for image classification are made up of 10 - 100 million learnable weights and are therefore inherently prone to overfitting. The complexity of the weight count can be seen as a function of the number of channels, the spatial extent of the input and the number of layers of the network. Due to the use of convolutional layers the scaling of weight complexity is usually linear with regards to the resolution dimensions, but remains quadratic with respect to the number of channels. Active research in recent years in terms of using multigrid inspired ideas in deep neural networks have shown that on one hand a significant number of weights can be saved by appropriate weight sharing and on the other that a hierarchical structure in the channel dimension can improve the weight complexity to linear. In this work, we combine these multigrid ideas to introduce a joint framework of multigrid inspired architectures, that exploit multigrid structures in all relevant dimensions to achieve linear weight complexity scaling and drastically reduced weight counts. Our experiments show that this structured reduction in weight count is able to reduce overfitting and thus shows improved performance over state-of-the-art ResNet architectures on typical image classification benchmarks at lower network complexity.

READ FULL TEXT
research
10/23/2022

Drastically Reducing the Number of Trainable Parameters in Deep CNNs by Inter-layer Kernel-sharing

Deep convolutional neural networks (DCNNs) have become the state-of-the-...
research
09/12/2019

Complexity-Scalable Neural Network Based MIMO Detection With Learnable Weight Scaling

This paper introduces a framework for systematic complexity scaling of d...
research
11/19/2016

Quantized neural network design under weight capacity constraint

The complexity of deep neural network algorithms for hardware implementa...
research
05/28/2019

RecNets: Channel-wise Recurrent Convolutional Neural Networks

In this paper, we introduce Channel-wise recurrent convolutional neural ...
research
01/11/2023

Deep Axial Hypercomplex Networks

Over the past decade, deep hypercomplex-inspired networks have enhanced ...
research
03/05/2023

Reparameterization through Spatial Gradient Scaling

Reparameterization aims to improve the generalization of deep neural net...
research
11/16/2018

Residual Convolutional Neural Network Revisited with Active Weighted Mapping

In visual recognition, the key to the performance improvement of ResNet ...

Please sign up or login with your details

Forgot password? Click here to reset