Tiled Squeeze-and-Excite: Channel Attention With Local Spatial Context

07/05/2021
by   Niv Vosco, et al.
0

In this paper we investigate the amount of spatial context required for channel attention. To this end we study the popular squeeze-and-excite (SE) block which is a simple and lightweight channel attention mechanism. SE blocks and its numerous variants commonly use global average pooling (GAP) to create a single descriptor for each channel. Here, we empirically analyze the amount of spatial context needed for effective channel attention and find that limited localcontext on the order of seven rows or columns of the original image is sufficient to match the performance of global context. We propose tiled squeeze-and-excite (TSE), which is a framework for building SE-like blocks that employ several descriptors per channel, with each descriptor based on local context only. We further show that TSE is a drop-in replacement for the SE block and can be used in existing SE networks without re-training. This implies that local context descriptors are similar both to each other and to the global context descriptor. Finally, we show that TSE has important practical implications for deployment of SE-networks to dataflow AI accelerators due to their reduced pipeline buffering requirements. For example, using TSE reduces the amount of activation pipeline buffering in EfficientDetD2 by 90 to SE (from 50M to 4.77M) without loss of accuracy. Our code and pre-trained models will be publicly available.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/06/2019

Linear Context Transform Block

Squeeze-and-Excitation (SE) block presents a channel attention mechanism...
research
01/06/2019

Channel Locality Block: A Variant of Squeeze-and-Excitation

Attention mechanism is a hot spot in deep learning field. Using channel ...
research
11/20/2018

Fading of collective attention shapes the evolution of linguistic variants

Language change involves the competition between alternative linguistic ...
research
10/18/2021

Abnormal Occupancy Grid Map Recognition using Attention Network

The occupancy grid map is a critical component of autonomous positioning...
research
09/05/2017

Squeeze-and-Excitation Networks

Convolutional neural networks are built upon the convolution operation, ...
research
04/22/2019

Stochastic Region Pooling: Make Attention More Expressive

Global Average Pooling (GAP) is used by default on the channel-wise atte...
research
05/22/2018

Global-and-local attention networks for visual recognition

State-of-the-art deep convolutional networks (DCNs) such as squeeze-and-...

Please sign up or login with your details

Forgot password? Click here to reset