"BNN - BN = ?": Training Binary Neural Networks without Batch Normalization

04/16/2021
by   Tianlong Chen, et al.
0

Batch normalization (BN) is a key facilitator and considered essential for state-of-the-art binary neural networks (BNN). However, the BN layer is costly to calculate and is typically implemented with non-binary parameters, leaving a hurdle for the efficient implementation of BNN training. It also introduces undesirable dependence between samples within each batch. Inspired by the latest advance on Batch Normalization Free (BN-Free) training, we extend their framework to training BNNs, and for the first time demonstrate that BNs can be completed removed from BNN training and inference regimes. By plugging in and customizing techniques including adaptive gradient clipping, scale weight standardization, and specialized bottleneck block, a BN-free BNN is capable of maintaining competitive accuracy compared to its BN-based counterpart. Extensive experiments validate the effectiveness of our proposal across diverse BNN backbones and datasets. For example, after removing BNs from the state-of-the-art ReActNets, it can still be trained with our proposed methodology to achieve 92.08 CIFAR-100, and ImageNet respectively, with marginal performance drop (0.23 available at: https://github.com/VITA-Group/BNN_NoBN.

READ FULL TEXT

page 1

page 7

research
12/23/2021

A Robust Initialization of Residual Blocks for Effective ResNet Training without Batch Normalization

Batch Normalization is an essential component of all state-of-the-art ne...
research
11/23/2020

Comparing Normalization Methods for Limited Batch Size Segmentation Neural Networks

The widespread use of Batch Normalization has enabled training deeper ne...
research
07/04/2022

Removing Batch Normalization Boosts Adversarial Training

Adversarial training (AT) defends deep neural networks against adversari...
research
01/01/2020

A Comprehensive and Modularized Statistical Framework for Gradient Norm Equality in Deep Neural Networks

In recent years, plenty of metrics have been proposed to identify networ...
research
02/22/2021

Sandwich Batch Normalization

We present Sandwich Batch Normalization (SaBN), an embarrassingly easy i...
research
08/21/2020

A constrained recursion algorithm for batch normalization of tree-sturctured LSTM

Tree-structured LSTM is promising way to consider long-distance interact...
research
03/22/2021

Adversarial Feature Augmentation and Normalization for Visual Recognition

Recent advances in computer vision take advantage of adversarial data au...

Please sign up or login with your details

Forgot password? Click here to reset