Evolutionary Optimization of Deep Learning Activation Functions

02/17/2020
by   Garrett Bingham, et al.
0

The choice of activation function can have a large effect on the performance of a neural network. While there have been some attempts to hand-engineer novel activation functions, the Rectified Linear Unit (ReLU) remains the most commonly-used in practice. This paper shows that evolutionary algorithms can discover novel activation functions that outperform ReLU. A tree-based search space of candidate activation functions is defined and explored with mutation, crossover, and exhaustive search. Experiments on training wide residual networks on the CIFAR-10 and CIFAR-100 image datasets show that this approach is effective. Replacing ReLU with evolved activation functions results in statistically significant increases in network accuracy. Optimal performance is achieved when evolution is allowed to customize activation functions to a particular task; however, these novel activation functions are shown to generalize, achieving high performance across tasks. Evolutionary optimization of activation functions is therefore a promising new dimension of metalearning in neural networks.

READ FULL TEXT
research
06/05/2020

Discovering Parametric Activation Functions

Recent studies have shown that the choice of activation function can sig...
research
05/12/2023

Saturated Non-Monotonic Activation Functions

Activation functions are essential to deep learning networks. Popular an...
research
06/24/2022

Evolution of Activation Functions for Deep Learning-Based Image Classification

Activation functions (AFs) play a pivotal role in the performance of neu...
research
08/29/2022

Normalized Activation Function: Toward Better Convergence

Activation functions are essential for neural networks to introduce non-...
research
05/30/2021

Evolution of Activation Functions: An Empirical Investigation

The hyper-parameters of a neural network are traditionally designed thro...
research
05/18/2023

Learning Activation Functions for Sparse Neural Networks

Sparse Neural Networks (SNNs) can potentially demonstrate similar perfor...
research
10/17/2020

Squashing activation functions in benchmark tests: towards eXplainable Artificial Intelligence using continuous-valued logic

Over the past few years, deep neural networks have shown excellent resul...

Please sign up or login with your details

Forgot password? Click here to reset