Attentive Normalization

08/04/2019
by   Xilai Li, et al.
0

Batch Normalization (BN) is a vital pillar in the development of deep learning with many recent variations such as Group Normalization (GN) and Switchable Normalization. Channel-wise feature attention methods such as the squeeze-and-excitation (SE) unit have also shown impressive performance improvement. BN and its variants take into account different ways of computing the mean and variance within a min-batch for feature normalization, followed by a learnable channel-wise affine transformation. SE explicitly learns how to adaptively recalibrate channel-wise feature responses. They have been studied separately, however. In this paper, we propose a novel and lightweight integration of feature normalization and feature channel-wise attention. We present Attentive Normalization (AN) as a simple and unified alternative. AN absorbs SE into the affine transformation of BN. AN learns a small number of scale and offset parameters per channel (i.e., different affine transformations). Their weighted sums (i.e., mixture) are used in the final affine transformation. The weights are instance-specific and learned in a way that channel-wise attention is considered, similar in spirit to the squeeze module in the SE unit. AN is complementary and applicable to existing variants of BN. In experiments, we test AN in the ImageNet-1K classification dataset and the MS-COCO object detection and instance segmentation dataset with significantly better performance obtained than the vanilla BN. Our AN also outperforms two state-of-the-art variants of BN, GN and SN. The source code will be released at <http://github.com/ivMCL/AttentiveNorm>.

READ FULL TEXT

page 2

page 5

research
09/25/2019

Gated Channel Transformation for Visual Recognition

In this work, we propose a generally applicable transformation unit for ...
research
07/31/2019

An Empirical Study of Batch Normalization and Group Normalization in Conditional Computation

Batch normalization has been widely used to improve optimization in deep...
research
10/26/2021

Revisiting Batch Normalization

Batch normalization (BN) is comprised of a normalization component follo...
research
01/15/2021

Dynamic Normalization

Batch Normalization has become one of the essential components in CNN. I...
research
05/25/2019

DIANet: Dense-and-Implicit Attention Network

Attention-based deep neural networks (DNNs) that emphasize the informati...
research
01/29/2022

Task-Balanced Batch Normalization for Exemplar-based Class-Incremental Learning

Batch Normalization (BN) is an essential layer for training neural netwo...
research
07/28/2023

AC-Norm: Effective Tuning for Medical Image Analysis via Affine Collaborative Normalization

Driven by the latest trend towards self-supervised learning (SSL), the p...

Please sign up or login with your details

Forgot password? Click here to reset