Normalizing the Normalizers: Comparing and Extending Network Normalization Schemes

11/14/2016
by   Mengye Ren, et al.
0

Normalization techniques have only recently begun to be exploited in supervised learning tasks. Batch normalization exploits mini-batch statistics to normalize the activations. This was shown to speed up training and result in better models. However its success has been very limited when dealing with recurrent neural networks. On the other hand, layer normalization normalizes the activations across all activities within a layer. This was shown to work well in the recurrent setting. In this paper we propose a unified view of normalization techniques, as forms of divisive normalization, which includes layer and batch normalization as special cases. Our second contribution is the finding that a small modification to these normalization schemes, in conjunction with a sparse regularizer on the activations, leads to significant benefits over standard normalization techniques. We demonstrate the effectiveness of our unified divisive normalization framework in the context of convolutional neural nets and recurrent neural networks, showing improvements over baselines in image classification, language modeling as well as super-resolution.

READ FULL TEXT

page 10

page 11

page 16

research
07/21/2016

Layer Normalization

Training state-of-the-art, deep neural networks is computationally expen...
research
05/15/2019

Online Normalization for Training Neural Networks

Online Normalization is a new technique for normalizing the hidden activ...
research
06/07/2021

Proxy-Normalizing Activations to Match Batch Normalization while Removing Batch Dependence

We investigate the reasons for the performance degradation incurred with...
research
08/05/2023

On problematic practice of using normalization in Self-modeling/Multivariate Curve Resolution (S/MCR)

The paper is briefly dealing with greater or lesser misused normalizatio...
research
06/10/2021

Beyond BatchNorm: Towards a General Understanding of Normalization in Deep Learning

Inspired by BatchNorm, there has been an explosion of normalization laye...
research
07/09/2019

Positional Normalization

A widely deployed method for reducing the training time of deep neural n...
research
02/19/2021

Convolutional Normalization

As the deep neural networks are being applied to complex tasks, the size...

Please sign up or login with your details

Forgot password? Click here to reset