Deductron - A Recurrent Neural Network

06/23/2018
by   Marek Rychlik, et al.
0

The current paper is a study in Recurrent Neural Networks (RNN), motivated by the lack of examples simple enough so that they can be thoroughly understood theoretically, but complex enough to be realistic. We constructed an example of structured data, motivated by problems from image-to-text conversion (OCR), which requires long-term memory to decode. Our data is a simple writing system, encoding characters 'X' and 'O' as their upper halves, which is possible due to symmetry of the two characters. The characters can be connected, as in some languages using cursive, such as Arabic (abjad). The string 'XOOXXO' may be encoded as '∨∧-1.5pt∧∨-1.5pt∨∧'. It follows that we may need to know arbitrarily long past to decode a current character, thus requiring long-term memory. Subsequently we constructed an RNN capable of decoding sequences encoded in this manner. Rather than by training, we constructed our RNN "by inspection", i.e. we guessed its weights. This involved a sequence of steps. We wrote a conventional program which decodes the sequences as the example above. Subsequently, we interpreted the program as a neural network (the only example of this kind known to us). Finally, we generalized this neural network to discover a new RNN architecture whose instance is our handcrafted RNN. It turns out to be a 3 layer network, where the middle layer is capable of performing simple logical inferences; thus the name "deductron". It is demonstrated that it is possible to train our network by simulated annealing. Also, known variants of stochastic gradient descent (SGD) methods are shown to work.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/14/2016

Recurrent neural network training with preconditioned stochastic gradient descent

This paper studies the performance of a recently proposed preconditioned...
research
04/18/2019

Language Modeling through Long Term Memory Network

Recurrent Neural Networks (RNN), Long Short-Term Memory Networks (LSTM),...
research
06/02/2021

Warming-up recurrent neural networks to maximize reachable multi-stability greatly improves learning

Training recurrent neural networks is known to be difficult when time de...
research
02/14/2014

A Clockwork RNN

Sequence prediction and classification are ubiquitous and challenging pr...
research
02/27/2020

Hierarchical Memory Decoding for Video Captioning

Recent advances of video captioning often employ a recurrent neural netw...
research
01/24/2018

PRNN: Recurrent Neural Network with Persistent Memory

Although Recurrent Neural Network (RNN) has been a powerful tool for mod...
research
04/16/2020

Investigating Efficient Learning and Compositionality in Generative LSTM Networks

When comparing human with artificial intelligence, one major difference ...

Please sign up or login with your details

Forgot password? Click here to reset