A New Activation Function for Training Deep Neural Networks to Avoid Local Minimum

05/15/2020
by   Abhinav Mishra, et al.
1

Activation functions have a major role to play and hence are very important while training neural networks. Understanding various activation functions and their advantages and disadvantages is crucial to achieve better results. This paper will first introduce common types of non linear activation functions and then evaluate their characteristics with their pros and cons. We will be focussing only on deep neural networks as it has proven to be much more difficult to train while avoiding overfitting. We have proposed a new activation function named - Abhinav which adds a non linearity with parametric coefficients to the Swish activation function. This has proven to give better results on the MNIST dataset. We reason this is because the model avoids getting stuck in local minima due to only the sigmoidal term present in the Swish function. The coefficients are automatically adjusted in each and every iteration i.e. coefficient values are reduced if the error is large and also sometimes reducing it to zero thus removing coefficients present in the polynomial. This new activation function can be made to generalize to other tasks including multi class and multi label image classification also.

READ FULL TEXT

page 3

page 7

research
07/31/2023

STL: A Signed and Truncated Logarithm Activation Function for Neural Networks

Activation functions play an essential role in neural networks. They pro...
research
03/14/2021

A Modified Batch Intrinsic Plasticity Method for Pre-training the Random Coefficients of Extreme Learning Machines

In extreme learning machines (ELM) the hidden-layer coefficients are ran...
research
09/30/2021

Introducing the DOME Activation Functions

In this paper, we introduce a novel non-linear activation function that ...
research
06/26/2018

Gradient Acceleration in Activation Functions

Dropout has been one of standard approaches to train deep neural network...
research
02/07/2020

Ill-Posedness and Optimization Geometry for Nonlinear Neural Network Training

In this work we analyze the role nonlinear activation functions play at ...
research
10/22/2021

Logical Activation Functions: Logit-space equivalents of Boolean Operators

Neuronal representations within artificial neural networks are commonly ...
research
05/01/2023

Activation Functions Not To Active: A Plausible Theory on Interpreting Neural Networks

Researchers commonly believe that neural networks model a high-dimension...

Please sign up or login with your details

Forgot password? Click here to reset