Neural Turing Machines: Convergence of Copy Tasks

12/07/2016
by   Janez Aleš, et al.
0

The architecture of neural Turing machines is differentiable end to end and is trainable with gradient descent methods. Due to their large unfolded depth Neural Turing Machines are hard to train and because of their linear access of complete memory they do not scale. Other architectures have been studied to overcome these difficulties. In this report we focus on improving the quality of prediction of the original linear memory architecture on copy and repeat copy tasks. Copy task predictions on sequences of length six times larger than those the neural Turing machine was trained on prove to be highly accurate and so do predictions of repeat copy tasks for sequences with twice the repetition number and twice the sequence length neural Turing machine was trained on.

READ FULL TEXT

page 2

page 3

page 4

page 5

research
10/20/2014

Neural Turing Machines

We extend the capabilities of neural networks by coupling them to extern...
research
10/12/2017

HyperENTM: Evolving Scalable Neural Turing Machines through HyperNEAT

Recent developments within memory-augmented neural networks have solved ...
research
10/14/2015

Structured Memory for Neural Turing Machines

Neural Turing Machines (NTM) contain memory component that simulates "wo...
research
06/30/2016

Dynamic Neural Turing Machine with Soft and Hard Addressing Schemes

We extend neural Turing machine (NTM) model into a dynamic neural Turing...
research
11/09/2016

Lie-Access Neural Turing Machines

External neural memory structures have recently become a popular tool fo...
research
03/18/2020

Progress Extrapolating Algorithmic Learning to Arbitrary Sequence Lengths

Recent neural network models for algorithmic tasks have led to significa...
research
04/04/2017

Using Echo State Networks for Cryptography

Echo state networks are simple recurrent neural networks that are easy t...

Please sign up or login with your details

Forgot password? Click here to reset