Empirical Evaluation of A New Approach to Simplifying Long Short-term Memory (LSTM)

12/12/2016
by   Yuzhen Lu, et al.
0

The standard LSTM, although it succeeds in the modeling long-range dependences, suffers from a highly complex structure that can be simplified through modifications to its gate units. This paper was to perform an empirical comparison between the standard LSTM and three new simplified variants that were obtained by eliminating input signal, bias and hidden unit signal from individual gates, on the tasks of modeling two sequence datasets. The experiments show that the three variants, with reduced parameters, can achieve comparable performance with the standard LSTM. Due attention should be paid to turning the learning rate to achieve high accuracies

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/12/2017

Simplified Gating in Long Short-term Memory (LSTM) Recurrent Neural Networks

The standard LSTM recurrent neural networks while very powerful in long-...
research
04/13/2018

The unreasonable effectiveness of the forget gate

Given the success of the gated recurrent unit, a natural question is whe...
research
06/08/2018

Towards Binary-Valued Gates for Robust LSTM Training

Long Short-Term Memory (LSTM) is one of the most widely used recurrent s...
research
01/18/2019

Slim LSTM networks: LSTM_6 and LSTM_C6

We have shown previously that our parameter-reduced variants of Long Sho...
research
03/13/2015

LSTM: A Search Space Odyssey

Several variants of the Long Short-Term Memory (LSTM) architecture for r...
research
10/16/2018

Reduced-Gate Convolutional LSTM Using Predictive Coding for Spatiotemporal Prediction

Spatiotemporal sequence prediction is an important problem in deep learn...
research
02/13/2019

Modeling default rate in P2P lending via LSTM

With the fast development of peer to peer (P2P) lending, financial insti...

Please sign up or login with your details

Forgot password? Click here to reset