Attention Based Pruning for Shift Networks

05/29/2019
by   Ghouthi Boukli Hacene, et al.
0

In many application domains such as computer vision, Convolutional Layers (CLs) are key to the accuracy of deep learning methods. However, it is often required to assemble a large number of CLs, each containing thousands of parameters, in order to reach state-of-the-art accuracy, thus resulting in complex and demanding systems that are poorly fitted to resource-limited devices. Recently, methods have been proposed to replace the generic convolution operator by the combination of a shift operation and a simpler 1x1 convolution. The resulting block, called Shift Layer (SL), is an efficient alternative to CLs in the sense it allows to reach similar accuracies on various tasks with faster computations and fewer parameters. In this contribution, we introduce Shift Attention Layers (SALs), which extend SLs by using an attention mechanism that learns which shifts are the best at the same time the network function is trained. We demonstrate SALs are able to outperform vanilla SLs (and CLs) on various object recognition benchmarks while significantly reducing the number of float operations and parameters for the inference.

READ FULL TEXT
research
12/17/2020

Attention-based Image Upsampling

Convolutional layers are an integral part of many deep neural network so...
research
05/30/2019

DeepShift: Towards Multiplication-Less Neural Networks

Deep learning models, especially DCNN have obtained high accuracies in s...
research
03/13/2019

All You Need is a Few Shifts: Designing Efficient Convolutional Neural Networks for Image Classification

Shift operation is an efficient alternative over depthwise separable con...
research
10/21/2019

CPWC: Contextual Point Wise Convolution for Object Recognition

Convolutional layers are a major driving force behind the successes of d...
research
11/08/2019

On the Relationship between Self-Attention and Convolutional Layers

Recent trends of incorporating attention mechanisms in vision have led r...
research
11/11/2022

Dual Complementary Dynamic Convolution for Image Recognition

As a powerful engine, vanilla convolution has promoted huge breakthrough...
research
04/02/2014

Exploiting Linear Structure Within Convolutional Networks for Efficient Evaluation

We present techniques for speeding up the test-time evaluation of large ...

Please sign up or login with your details

Forgot password? Click here to reset