Stochastic Region Pooling: Make Attention More Expressive

04/22/2019
by   Mingnan Luo, et al.
0

Global Average Pooling (GAP) is used by default on the channel-wise attention mechanism to extract channel descriptors. However, the simple global aggregation method of GAP is easy to make the channel descriptors have homogeneity, which weakens the detail distinction between feature maps, thus affecting the performance of the attention mechanism. In this work, we propose a novel method for channel-wise attention network, called Stochastic Region Pooling (SRP), which makes the channel descriptors more representative and diversity by encouraging the feature map to have more or wider important feature responses. Also, SRP is the general method for the attention mechanisms without any additional parameters or computation. It can be widely applied to attention networks without modifying the network structure. Experimental results on image recognition datasets including CIAFR-10/100, ImageNet and three Fine-grained datasets (CUB-200-2011, Stanford Cars and Stanford Dogs) show that SRP brings the significant improvements of the performance over efficient CNNs and achieves the state-of-the-art results.

READ FULL TEXT

page 2

page 4

page 6

page 7

page 8

page 11

research
10/17/2022

Cross-layer Attention Network for Fine-grained Visual Categorization

Learning discriminative representations for subtle localized details pla...
research
12/22/2020

FcaNet: Frequency Channel Attention Networks

Attention mechanism, especially channel attention, has gained great succ...
research
08/25/2023

MMBAttn: Max-Mean and Bit-wise Attention for CTR Prediction

With the increasing complexity and scale of click-through rate (CTR) pre...
research
12/13/2022

CAT: Learning to Collaborate Channel and Spatial Attention from Multi-Information Fusion

Channel and spatial attention mechanism has proven to provide an evident...
research
03/04/2021

Coordinate Attention for Efficient Mobile Network Design

Recent studies on mobile network design have demonstrated the remarkable...
research
10/07/2020

Channel Recurrent Attention Networks for Video Pedestrian Retrieval

Full attention, which generates an attention value per element of the in...
research
07/05/2021

Tiled Squeeze-and-Excite: Channel Attention With Local Spatial Context

In this paper we investigate the amount of spatial context required for ...

Please sign up or login with your details

Forgot password? Click here to reset