Rethinking "Batch" in BatchNorm

05/17/2021
by   Yuxin Wu, et al.
0

BatchNorm is a critical building block in modern convolutional neural networks. Its unique property of operating on "batches" instead of individual samples introduces significantly different behaviors from most other operations in deep learning. As a result, it leads to many hidden caveats that can negatively impact model's performance in subtle ways. This paper thoroughly reviews such problems in visual recognition tasks, and shows that a key to address them is to rethink different choices in the concept of "batch" in BatchNorm. By presenting these caveats and their mitigations, we hope this review can help researchers use BatchNorm more effectively.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/05/2015

Batch Normalized Recurrent Neural Networks

Recurrent Neural Networks (RNNs) are powerful models for sequential data...
research
10/18/2016

Master's Thesis : Deep Learning for Visual Recognition

The goal of our research is to develop methods advancing automatic visua...
research
05/20/2022

Kernel Normalized Convolutional Networks

Existing deep convolutional neural network (CNN) architectures frequentl...
research
08/03/2018

Large Scale Language Modeling: Converging on 40GB of Text in Four Hours

Recent work has shown how to train Convolutional Neural Networks (CNNs) ...
research
01/19/2023

Batch Prompting: Efficient Inference with Large Language Model APIs

Performing inference on hundreds of thousands of samples with large lang...
research
08/04/2022

Beer2Vec : Extracting Flavors from Reviews for Thirst-Quenching Recommandations

This paper introduces the Beer2Vec model that allows the most popular al...

Please sign up or login with your details

Forgot password? Click here to reset