TanhExp: A Smooth Activation Function with High Convergence Speed for Lightweight Neural Networks

03/22/2020
by   Xinyu Liu, et al.
9

Lightweight or mobile neural networks used for real-time computer vision tasks contain fewer parameters than normal networks, which lead to a constrained performance. In this work, we proposed a novel activation function named Tanh Exponential Activation Function (TanhExp) which can improve the performance for these networks on image classification task significantly. The definition of TanhExp is f(x) = xtanh(e^x). We demonstrate the simplicity, efficiency, and robustness of TanhExp on various datasets and network models and TanhExp outperforms its counterparts in both convergence speed and accuracy. Its behaviour also remains stable even with noise added and dataset altered. We show that without increasing the size of the network, the capacity of lightweight neural networks can be enhanced by TanhExp with only a few training epochs and no extra parameters added.

READ FULL TEXT

page 1

page 2

page 4

research
01/15/2023

Empirical study of the modulus as activation function in computer vision applications

In this work we propose a new non-monotonic activation function: the mod...
research
01/14/2021

A Multiple Classifier Approach for Concatenate-Designed Neural Networks

This article introduces a multiple classifier method to improve the perf...
research
09/03/2021

Using Topological Framework for the Design of Activation Function and Model Pruning in Deep Neural Networks

Success of deep neural networks in diverse tasks across domains of compu...
research
08/09/2023

TSSR: A Truncated and Signed Square Root Activation Function for Neural Networks

Activation functions are essential components of neural networks. In thi...
research
08/30/2022

CAIR: Fast and Lightweight Multi-Scale Color Attention Network for Instagram Filter Removal

Image restoration is an important and challenging task in computer visio...
research
07/23/2020

Nonclosedness of the Set of Neural Networks in Sobolev Space

We examine the closedness of the set of realized neural networks of a fi...
research
10/08/2020

Randomized Overdrive Neural Networks

By processing audio signals in the time-domain with randomly weighted te...

Please sign up or login with your details

Forgot password? Click here to reset