LSTM Networks Can Perform Dynamic Counting

06/09/2019
by   Mirac Suzgun, et al.
0

In this paper, we systematically assess the ability of standard recurrent networks to perform dynamic counting and to encode hierarchical representations. All the neural models in our experiments are designed to be small-sized networks both to prevent them from memorizing the training sets and to visualize and interpret their behaviour at test time. Our results demonstrate that the Long Short-Term Memory (LSTM) networks can learn to recognize the well-balanced parenthesis language (Dyck-1) and the shuffles of multiple Dyck-1 languages, each defined over different parenthesis-pairs, by emulating simple real-time k-counter machines. To the best of our knowledge, this work is the first study to introduce the shuffle languages to analyze the computational power of neural networks. We also show that a single-layer LSTM with only one hidden unit is practically sufficient for recognizing the Dyck-1 language. However, none of our recurrent networks was able to yield a good performance on the Dyck-2 language learning task, which requires a model to have a stack-like mechanism for recognition.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/23/2022

Image Classification using Sequence of Pixels

This study compares sequential image classification methods based on rec...
research
05/01/2018

A Taxonomy for Neural Memory Networks

In this paper, a taxonomy for memory networks is proposed based on their...
research
07/09/2019

Improving the Performance of the LSTM and HMM Models via Hybridization

Language models based on deep neural neural networks and traditionalstoc...
research
11/18/2016

Increasing the Interpretability of Recurrent Neural Networks Using Hidden Markov Models

As deep neural networks continue to revolutionize various application do...
research
10/09/2020

Recurrent babbling: evaluating the acquisition of grammar from limited input data

Recurrent Neural Networks (RNNs) have been shown to capture various aspe...
research
06/04/2019

Sequential Neural Networks as Automata

This work attempts to explain the types of computation that neural netwo...
research
02/10/2018

An LSTM Recurrent Network for Step Counting

Smartphones with sensors such as accelerometer and gyroscope can be used...

Please sign up or login with your details

Forgot password? Click here to reset