Recurrent Memory Array Structures

07/11/2016
by   Kamil Rocki, et al.
0

The following report introduces ideas augmenting standard Long Short Term Memory (LSTM) architecture with multiple memory cells per hidden unit in order to improve its generalization capabilities. It considers both deterministic and stochastic variants of memory operation. It is shown that the nondeterministic Array-LSTM approach improves state-of-the-art performance on character level text prediction achieving 1.402 BPC on enwik8 dataset. Furthermore, this report estabilishes baseline neural-based results of 1.12 BPC and 1.19 BPC for enwik9 and enwik10 datasets respectively.

READ FULL TEXT
research
01/02/2019

Performance of Three Slim Variants of The Long Short-Term Memory (LSTM) Layer

The Long Short-Term Memory (LSTM) layer is an important advancement in t...
research
05/11/2021

Restoring Hebrew Diacritics Without a Dictionary

We demonstrate that it is feasible to diacritize Hebrew script without a...
research
10/03/2018

TWA – Ticket Locks Augmented with a Waiting Array

The classic ticket lock consists of ticket and grant fields. Arriving th...
research
11/14/2017

SkipFlow: Incorporating Neural Coherence Features for End-to-End Automatic Text Scoring

Deep learning has demonstrated tremendous potential for Automatic Text S...
research
02/08/2017

Character-level Deep Conflation for Business Data Analytics

Connecting different text attributes associated with the same entity (co...
research
09/23/2019

Field typing for improved recognition on heterogeneous handwritten forms

Offline handwriting recognition has undergone continuous progress over t...
research
01/31/2020

Compensation of Fiber Nonlinearities in Digital Coherent Systems Leveraging Long Short-Term Memory Neural Networks

We introduce for the first time the utilization of Long short-term memor...

Please sign up or login with your details

Forgot password? Click here to reset