Nested LSTMs

01/31/2018
by   Joel Ruben Antony Moniz, et al.
0

We propose Nested LSTMs (NLSTM), a novel RNN architecture with multiple levels of memory. Nested LSTMs add depth to LSTMs via nesting as opposed to stacking. The value of a memory cell in an NLSTM is computed by an LSTM cell, which has its own inner memory cell. Specifically, instead of computing the value of the (outer) memory cell as c^outer_t = f_t c_t-1 + i_t g_t, NLSTM memory cells use the concatenation (f_t c_t-1, i_t g_t) as input to an inner LSTM (or NLSTM) memory cell, and set c^outer_t = h^inner_t. Nested LSTMs outperform both stacked and single-layer LSTMs with similar numbers of parameters in our experiments on various character-level language modeling tasks, and the inner memories of an LSTM learn longer term dependencies compared with the higher-level units of a stacked LSTM.

READ FULL TEXT
research
09/07/2018

Cell-aware Stacked LSTMs for Modeling Sentences

We propose a method of stacking multiple long short-term memory (LSTM) l...
research
11/15/2018

Multi-cell LSTM Based Neural Language Model

Language models, being at the heart of many NLP problems, are always of ...
research
08/16/2015

Depth-Gated LSTM

In this short note, we present an extension of long short-term memory (L...
research
03/19/2019

IndyLSTMs: Independently Recurrent LSTMs

We introduce Independently Recurrent Long Short-term Memory cells: IndyL...
research
11/30/2019

Modeling German Verb Argument Structures: LSTMs vs. Humans

LSTMs have proven very successful at language modeling. However, it rema...
research
04/22/2022

Nested Papercrafts for Anatomical and Biological Edutainment

In this paper, we present a new workflow for the computer-aided generati...
research
08/04/2018

MCRM: Mother Compact Recurrent Memory A Biologically Inspired Recurrent Neural Network Architecture

LSTMs and GRUs are the most common recurrent neural network architecture...

Please sign up or login with your details

Forgot password? Click here to reset