Deep Learning with S-shaped Rectified Linear Activation Units

12/22/2015
by   Xiaojie Jin, et al.
0

Rectified linear activation units are important components for state-of-the-art deep convolutional networks. In this paper, we propose a novel S-shaped rectified linear activation unit (SReLU) to learn both convex and non-convex functions, imitating the multiple function forms given by the two fundamental laws, namely the Webner-Fechner law and the Stevens law, in psychophysics and neural sciences. Specifically, SReLU consists of three piecewise linear functions, which are formulated by four learnable parameters. The SReLU is learned jointly with the training of the whole deep network through back propagation. During the training phase, to initialize SReLU in different layers, we propose a "freezing" method to degenerate SReLU into a predefined leaky rectified linear unit in the initial several training epochs and then adaptively learn the good initial values. SReLU can be universally used in the existing deep networks with negligible additional parameters and computation cost. Experiments with two popular CNN architectures, Network in Network and GoogLeNet on scale-various benchmarks including CIFAR10, CIFAR100, MNIST and ImageNet demonstrate that SReLU achieves remarkable improvement compared to other activation functions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/05/2021

An Analysis of State-of-the-art Activation Functions For Supervised Deep Neural Network

This paper provides an analysis of state-of-the-art activation functions...
research
01/25/2021

Parametric Rectified Power Sigmoid Units: Learning Nonlinear Neural Transfer Analytical Forms

The paper proposes representation functionals in a dual paradigm where l...
research
08/03/2015

On the Importance of Normalisation Layers in Deep Learning with Piecewise Linear Activation Units

Deep feedforward neural networks with piecewise linear activations are c...
research
07/15/2020

Attention as Activation

Activation functions and attention mechanisms are typically treated as h...
research
01/05/2020

Cooperative Initialization based Deep Neural Network Training

Researchers have proposed various activation functions. These activation...
research
07/07/2019

Towards Robust, Locally Linear Deep Networks

Deep networks realize complex mappings that are often understood by thei...
research
11/29/2021

First Power Linear Unit with Sign

This paper proposes a novel and insightful activation method termed FPLU...

Please sign up or login with your details

Forgot password? Click here to reset