The Quest for the Golden Activation Function

08/02/2018
by   Mina Basirat, et al.
0

Deep Neural Networks have been shown to be beneficial for a variety of tasks, in particular allowing for end-to-end learning and reducing the requirement for manual design decisions. However, still many parameters have to be chosen in advance, also raising the need to optimize them. One important, but often ignored system parameter is the selection of a proper activation function. Thus, in this paper we target to demonstrate the importance of activation functions in general and show that for different tasks different activation functions might be meaningful. To avoid the manual design or selection of activation functions, we build on the idea of genetic algorithms to learn the best activation function for a given task. In addition, we introduce two new activation functions, ELiSH and HardELiSH, which can easily be incorporated in our framework. In this way, we demonstrate for three different image classification benchmarks that different activation functions are learned, also showing improved results compared to typically used baselines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/13/2023

Deep Network Approximation: Beyond ReLU to Diverse Activation Functions

This paper explores the expressive power of deep neural networks for a d...
research
08/08/2023

Learning Specialized Activation Functions for Physics-informed Neural Networks

Physics-informed neural networks (PINNs) are known to suffer from optimi...
research
08/30/2022

Transformers with Learnable Activation Functions

Activation functions can have a significant impact on reducing the topol...
research
03/16/2022

Adaptive n-ary Activation Functions for Probabilistic Boolean Logic

Balancing model complexity against the information contained in observed...
research
06/09/2020

A Note on Deepfake Detection with Low-Resources

Deepfakes are videos that include changes, quite often substituting face...
research
05/25/2023

Embeddings between Barron spaces with higher order activation functions

The approximation properties of infinitely wide shallow neural networks ...
research
01/13/2021

Reproducing Activation Function for Deep Learning

In this paper, we propose the reproducing activation function to improve...

Please sign up or login with your details

Forgot password? Click here to reset