Depth-Gated LSTM

08/16/2015
by   Kaisheng Yao, et al.
0

In this short note, we present an extension of long short-term memory (LSTM) neural networks to using a depth gate to connect memory cells of adjacent layers. Doing so introduces a linear dependence between lower and upper layer recurrent units. Importantly, the linear dependence is gated through a gating function, which we call depth gate. This gate is a function of the lower layer memory cell, the input to and the past memory cell of this layer. We conducted experiments and verified that this new architecture of LSTMs was able to improve machine translation and language modeling performances.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/02/2019

Performance of Three Slim Variants of The Long Short-Term Memory (LSTM) Layer

The Long Short-Term Memory (LSTM) layer is an important advancement in t...
research
12/24/2021

Self-Gated Memory Recurrent Network for Efficient Scalable HDR Deghosting

We propose a novel recurrent network-based HDR deghosting method for fus...
research
09/07/2018

Cell-aware Stacked LSTMs for Modeling Sentences

We propose a method of stacking multiple long short-term memory (LSTM) l...
research
01/31/2018

Nested LSTMs

We propose Nested LSTMs (NLSTM), a novel RNN architecture with multiple ...
research
07/05/2021

A comparison of LSTM and GRU networks for learning symbolic sequences

We explore relations between the hyper-parameters of a recurrent neural ...
research
03/19/2019

IndyLSTMs: Independently Recurrent LSTMs

We introduce Independently Recurrent Long Short-term Memory cells: IndyL...
research
08/04/2018

MCRM: Mother Compact Recurrent Memory

LSTMs and GRUs are the most common recurrent neural network architecture...

Please sign up or login with your details

Forgot password? Click here to reset