Minimum Description Length Recurrent Neural Networks

10/31/2021
by   Nur Lan, et al.
0

We train neural networks to optimize a Minimum Description Length score, i.e., to balance between the complexity of the network and its accuracy at a task. We show that networks trained with this objective function master tasks involving memory challenges such as counting, including cases that go beyond context-free languages. These learners master grammars for, e.g., a^nb^n, a^nb^nc^n, a^nb^2n, and a^nb^mc^n+m, and they perform addition. They do so with 100 also small and their inner workings are transparent. We thus provide formal proofs that their perfect accuracy holds not only on a given test set, but for any input sequence.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/16/2023

Benchmarking Neural Network Generalization for Grammar Induction

How well do neural networks generalize? Even for grammar induction tasks...
research
06/03/2019

A detailed study of recurrent neural networks used to model tasks in the cerebral cortex

We studied the properties of simple recurrent neural networks trained to...
research
02/14/2023

Interpolation Learning With Minimum Description Length

We prove that the Minimum Description Length learning rule exhibits temp...
research
05/24/2019

Neuro-Optimization: Learning Objective Functions Using Neural Networks

Mathematical optimization is widely used in various research fields. Wit...
research
06/18/2020

Stability of Internal States in Recurrent Neural Networks Trained on Regular Languages

We provide an empirical study of the stability of recurrent neural netwo...
research
10/09/2017

full-FORCE: A Target-Based Method for Training Recurrent Networks

Trained recurrent networks are powerful tools for modeling dynamic neura...
research
05/02/2023

Memory of recurrent networks: Do we compute it right?

Numerical evaluations of the memory capacity (MC) of recurrent neural ne...

Please sign up or login with your details

Forgot password? Click here to reset