ShuffleBlock: Shuffle to Regularize Deep Convolutional Neural Networks

06/17/2021
by   Sudhakar Kumawat, et al.
14

Deep neural networks have enormous representational power which leads them to overfit on most datasets. Thus, regularizing them is important in order to reduce overfitting and enhance their generalization capabilities. Recently, channel shuffle operation has been introduced for mixing channels in group convolutions in resource efficient networks in order to reduce memory and computations. This paper studies the operation of channel shuffle as a regularization technique in deep convolutional networks. We show that while random shuffling of channels during training drastically reduce their performance, however, randomly shuffling small patches between channels significantly improves their performance. The patches to be shuffled are picked from the same spatial locations in the feature maps such that a patch, when transferred from one channel to another, acts as structured noise for the later channel. We call this method "ShuffleBlock". The proposed ShuffleBlock module is easy to implement and improves the performance of several baseline networks on the task of image classification on CIFAR and ImageNet datasets. It also achieves comparable and in many cases better performance than many other regularization methods. We provide several ablation studies on selecting various hyperparameters of the ShuffleBlock module and propose a new scheduling method that further enhances its performance.

READ FULL TEXT

page 3

page 7

page 13

research
12/10/2021

A Discriminative Channel Diversification Network for Image Classification

Channel attention mechanisms in convolutional neural networks have been ...
research
03/20/2023

Parameter-Free Channel Attention for Image Classification and Super-Resolution

The channel attention mechanism is a useful technique widely employed in...
research
07/22/2017

PatchShuffle Regularization

This paper focuses on regularizing the training of the convolutional neu...
research
01/11/2023

Enhancing ResNet Image Classification Performance by using Parameterized Hypercomplex Multiplication

Recently, many deep networks have introduced hypercomplex and related ca...
research
08/21/2019

RBCN: Rectified Binary Convolutional Networks for Enhancing the Performance of 1-bit DCNNs

Binarized convolutional neural networks (BCNNs) are widely used to impro...
research
08/25/2023

Squeeze aggregated excitation network

Convolutional neural networks have spatial representations which read pa...
research
03/10/2021

MixMo: Mixing Multiple Inputs for Multiple Outputs via Deep Subnetworks

Recent strategies achieved ensembling "for free" by fitting concurrently...

Please sign up or login with your details

Forgot password? Click here to reset