Effects of the Nonlinearity in Activation Functions on the Performance of Deep Learning Models

10/14/2020
by   Nalinda Kulathunga, et al.
0

The nonlinearity of activation functions used in deep learning models are crucial for the success of predictive models. There are several commonly used simple nonlinear functions, including Rectified Linear Unit (ReLU) and Leaky-ReLU (L-ReLU). In practice, these functions remarkably enhance the model accuracy. However, there is limited insight into the functionality of these nonlinear activation functions in terms of why certain models perform better than others. Here, we investigate the model performance when using ReLU or L-ReLU as activation functions in different model architectures and data domains. Interestingly, we found that the application of L-ReLU is mostly effective when the number of trainable parameters in a model is relatively small. Furthermore, we found that the image classification models seem to perform well with L-ReLU in fully connected layers, especially when pre-trained models such as the VGG-16 are used for the transfer learning.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/05/2020

Discovering Parametric Activation Functions

Recent studies have shown that the choice of activation function can sig...
research
02/26/2021

Neural Generalization of Multiple Kernel Learning

Multiple Kernel Learning is a conventional way to learn the kernel funct...
research
09/06/2018

ANS: Adaptive Network Scaling for Deep Rectifier Reinforcement Learning Models

This work provides a thorough study on how reward scaling can affect per...
research
02/27/2023

Complex Clipping for Improved Generalization in Machine Learning

For many machine learning applications, a common input representation is...
research
05/04/2022

Convolutional and Residual Networks Provably Contain Lottery Tickets

The Lottery Ticket Hypothesis continues to have a profound practical imp...
research
12/17/2021

Adaptively Customizing Activation Functions for Various Layers

To enhance the nonlinearity of neural networks and increase their mappin...
research
04/25/2020

Compromise-free Bayesian neural networks

We conduct a thorough analysis of the relationship between the out-of-sa...

Please sign up or login with your details

Forgot password? Click here to reset