Phish: A Novel Hyper-Optimizable Activation Function

01/15/2022
by   Philip Naveen, et al.
0

Deep-learning models estimate values using backpropagation. The activation function within hidden layers is a critical component to minimizing loss in deep neural-networks. Rectified Linear (ReLU) has been the dominant activation function for the past decade. Swish and Mish are newer activation functions that have shown to yield better results than ReLU given specific circumstances. Phish is a novel activation function proposed here. It is a composite function defined as f(x) = xTanH(GELU(x)), where no discontinuities are apparent in the differentiated graph on the domain observed. Generalized networks were constructed using different activation functions. SoftMax was the output function. Using images from MNIST and CIFAR-10 databanks, these networks were trained to minimize sparse categorical crossentropy. A large scale cross-validation was simulated using stochastic Markov chains to account for the law of large numbers for the probability values. Statistical tests support the research hypothesis stating Phish could outperform other activation functions in classification. Future experiments would involve testing Phish in unsupervised learning algorithms and comparing it to more activation functions.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/02/2023

ErfReLU: Adaptive Activation Function for Deep Neural Network

Recent research has found that the activation function (AF) selected for...
research
12/15/2018

Flatten-T Swish: a thresholded ReLU-Swish-like activation function for deep learning

Activation functions are essential for deep learning methods to learn an...
research
05/20/2023

GELU Activation Function in Deep Learning: A Comprehensive Mathematical Analysis and Performance

Selecting the most suitable activation function is a critical factor in ...
research
10/30/2019

Sparsely Activated Networks: A new method for decomposing and compressing data

Recent literature on unsupervised learning focused on designing structur...
research
07/12/2019

Sparsely Activated Networks

Previous literature on unsupervised learning focused on designing struct...
research
05/26/2019

ProbAct: A Probabilistic Activation Function for Deep Neural Networks

Activation functions play an important role in the training of artificia...
research
10/21/2022

Stochastic Adaptive Activation Function

The simulation of human neurons and neurotransmission mechanisms has bee...

Please sign up or login with your details

Forgot password? Click here to reset