A Comprehensive Survey and Performance Analysis of Activation Functions in Deep Learning

09/29/2021
by   Shiv Ram Dubey, et al.
40

Neural networks have shown tremendous growth in recent years to solve numerous problems. Various types of neural networks have been introduced to deal with different types of problems. However, the main goal of any neural network is to transform the non-linearly separable input data into more linearly separable abstract features using a hierarchy of layers. These layers are combinations of linear and nonlinear functions. The most popular and common non-linearity layers are activation functions (AFs), such as Logistic Sigmoid, Tanh, ReLU, ELU, Swish and Mish. In this paper, a comprehensive overview and survey is presented for AFs in neural networks for deep learning. Different classes of AFs such as Logistic Sigmoid and Tanh based, ReLU based, ELU based, and Learning based are covered. Several characteristics of AFs such as output range, monotonicity, and smoothness are also pointed out. A performance comparison is also performed among 18 state-of-the-art AFs with different networks on different types of data. The insights of AFs are presented to benefit the researchers for doing further research and practitioners to select among different choices. The code used for experimental comparison is released at: <https://github.com/shivram1987/ActivationFunctions>.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/25/2021

Activation Functions in Artificial Neural Networks: A Systematic Overview

Activation functions shape the outputs of artificial neurons and, theref...
research
10/19/2022

A new activation for neural networks and its approximation

Deep learning with deep neural networks (DNNs) has attracted tremendous ...
research
07/08/2018

Separability is not the best goal for machine learning

Neural networks use their hidden layers to transform input data into lin...
research
02/11/2020

Goldilocks Neural Networks

We introduce the new "Goldilocks" class of activation functions, which n...
research
03/29/2021

Comparison of different convolutional neural network activation functions and methods for building ensembles

Recently, much attention has been devoted to finding highly efficient an...
research
03/22/2022

Exploring Linear Feature Disentanglement For Neural Networks

Non-linear activation functions, e.g., Sigmoid, ReLU, and Tanh, have ach...
research
12/17/2021

Adaptively Customizing Activation Functions for Various Layers

To enhance the nonlinearity of neural networks and increase their mappin...

Please sign up or login with your details

Forgot password? Click here to reset