An Investigation on Deep Learning with Beta Stabilizer

07/31/2020
by   Qi Liu, et al.
0

Artificial neural networks (ANN) have been used in many applications such like handwriting recognition and speech recognition. It is well-known that learning rate is a crucial value in the training procedure for artificial neural networks. It is shown that the initial value of learning rate can confoundedly affect the final result and this value is always set manually in practice. A new parameter called beta stabilizer has been introduced to reduce the sensitivity of the initial learning rate. But this method has only been proposed for deep neural network (DNN) with sigmoid activation function. In this paper we extended beta stabilizer to long short-term memory (LSTM) and investigated the effects of beta stabilizer parameters on different models, including LSTM and DNN with relu activation function. It is concluded that beta stabilizer parameters can reduce the sensitivity of learning rate with almost the same performance on DNN with relu activation function and LSTM. However, it is shown that the effects of beta stabilizer on DNN with relu activation function and LSTM are fewer than the effects on DNN with sigmoid activation function.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/06/2018

SBAF: A New Activation Function for Artificial Neural Net based Habitability Classification

We explore the efficacy of using a novel activation function in Artifici...
research
11/12/2020

Empirical Performance Analysis of Conventional Deep Learning Models for Recognition of Objects in 2-D Images

Artificial Neural Networks, an essential part of Deep Learning, are deri...
research
12/23/2019

Learn-able parameter guided Activation Functions

In this paper, we explore the concept of adding learn-able slope and mea...
research
05/26/2019

ProbAct: A Probabilistic Activation Function for Deep Neural Networks

Activation functions play an important role in the training of artificia...
research
03/31/2017

Noisy Softplus: an activation function that enables SNNs to be trained as ANNs

We extended the work of proposed activation function, Noisy Softplus, to...
research
10/16/2022

Stability of Accuracy for the Training of DNNs Via the Uniform Doubling Condition

We study the stability of accuracy for the training of deep neural netwo...
research
11/09/2017

Feed Forward and Backward Run in Deep Convolution Neural Network

Convolution Neural Networks (CNN), known as ConvNets are widely used in ...

Please sign up or login with your details

Forgot password? Click here to reset