Evolving Normalization-Activation Layers

04/06/2020
by   Hanxiao Liu, et al.
9

Normalization layers and activation functions are critical components in deep neural networks that frequently co-locate with each other. Instead of designing them separately, we unify them into a single computation graph, and evolve its structure starting from low-level primitives. Our layer search algorithm leads to the discovery of EvoNorms, a set of new normalization-activation layers that go beyond existing design patterns. Several of these layers enjoy the property of being independent from the batch statistics. Our experiments show that EvoNorms not only excel on a variety of image classification models including ResNets, MobileNets and EfficientNets, but also transfer well to Mask R-CNN for instance segmentation and BigGAN for image synthesis, outperforming BatchNorm and GroupNorm based layers by a significant margin in many cases.

READ FULL TEXT
research
09/12/2017

Shifting Mean Activation Towards Zero with Bipolar Activation Functions

We propose a simple extension to the ReLU-family of activation functions...
research
12/12/2019

Local Context Normalization: Revisiting Local Normalization

Normalization layers have been shown to improve convergence in deep neur...
research
06/08/2023

Layer-level activation mechanism

In this work, we propose a novel activation mechanism aimed at establish...
research
03/28/2022

To Fold or Not to Fold: a Necessary and Sufficient Condition on Batch-Normalization Layers Folding

Batch-Normalization (BN) layers have become fundamental components in th...
research
05/28/2023

On the impact of activation and normalization in obtaining isometric embeddings at initialization

In this paper, we explore the structure of the penultimate Gram matrix i...
research
06/08/2023

Normalization-Equivariant Neural Networks with Application to Image Denoising

In many information processing systems, it may be desirable to ensure th...
research
03/15/2022

Deep Learning without Shortcuts: Shaping the Kernel with Tailored Rectifiers

Training very deep neural networks is still an extremely challenging tas...

Please sign up or login with your details

Forgot password? Click here to reset