A Robust Initialization of Residual Blocks for Effective ResNet Training without Batch Normalization

12/23/2021
by   Enrico Civitelli, et al.
0

Batch Normalization is an essential component of all state-of-the-art neural networks architectures. However, since it introduces many practical issues, much recent research has been devoted to designing normalization-free architectures. In this paper, we show that weights initialization is key to train ResNet-like normalization-free networks. In particular, we propose a slight modification to the summation operation of a block output to the skip connection branch, so that the whole network is correctly initialized. We show that this modified architecture achieves competitive results on CIFAR-10 without further regularization nor algorithmic modifications.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/24/2020

Batch Normalization Biases Deep Residual Networks Towards Shallow Paths

Batch normalization has multiple benefits. It improves the conditioning ...
research
09/09/2017

Deep Residual Networks and Weight Initialization

Residual Network (ResNet) is the state-of-the-art architecture that real...
research
04/16/2021

"BNN - BN = ?": Training Binary Neural Networks without Batch Normalization

Batch normalization (BN) is a key facilitator and considered essential f...
research
02/27/2019

Equi-normalization of Neural Networks

Modern neural networks are over-parametrized. In particular, each rectif...
research
05/21/2017

Shake-Shake regularization

The method introduced in this paper aims at helping deep learning practi...
research
10/03/2022

Feature Embedding by Template Matching as a ResNet Block

Convolution blocks serve as local feature extractors and are the key to ...
research
10/05/2022

Dynamical Isometry for Residual Networks

The training success, training speed and generalization ability of neura...

Please sign up or login with your details

Forgot password? Click here to reset