Approximating Activation Functions

01/17/2020
by   Nicholas Gerard Timmons, et al.
13

ReLU is widely seen as the default choice for activation functions in neural networks. However, there are cases where more complicated functions are required. In particular, recurrent neural networks (such as LSTMs) make extensive use of both hyperbolic tangent and sigmoid functions. These functions are expensive to compute. We used function approximation techniques to develop replacements for these functions and evaluated them empirically on three popular network configurations. We find safe approximations that yield a 10 37 suitable for all cases we considered and we believe are appropriate replacements for all networks using these activation functions. We also develop ranged approximations which only apply in some cases due to restrictions on their input domain. Our ranged approximations yield a performance improvement of 20 considerably out perform the ad-hoc approximations used in Theano and the implementation of Word2Vec.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/25/2021

Activation Functions in Artificial Neural Networks: A Systematic Overview

Activation functions shape the outputs of artificial neurons and, theref...
research
07/02/2019

Best k-layer neural network approximations

We investigate the geometry of the empirical risk minimization problem f...
research
06/07/2013

Non-constant bounded holomorphic functions of hyperbolic numbers - Candidates for hyperbolic activation functions

The Liouville theorem states that bounded holomorphic complex functions ...
research
04/08/2020

The Loss Surfaces of Neural Networks with General Activation Functions

We present results extending the foundational work of Choromanska et al ...
research
06/22/2020

Advantages of biologically-inspired adaptive neural activation in RNNs during learning

Dynamic adaptation in single-neuron response plays a fundamental role in...
research
11/21/2022

DualApp: Tight Over-Approximation for Neural Network Robustness Verification via Under-Approximation

The robustness of neural networks is fundamental to the hosting system's...
research
02/11/2020

Goldilocks Neural Networks

We introduce the new "Goldilocks" class of activation functions, which n...

Please sign up or login with your details

Forgot password? Click here to reset