Provably Stable Interpretable Encodings of Context Free Grammars in RNNs with a Differentiable Stack

06/05/2020
by   Dr. John Stogin, et al.
0

Given a collection of strings belonging to a context free grammar (CFG) and another collection of strings not belonging to the CFG, how might one infer the grammar? This is the problem of grammatical inference. Since CFGs are the languages recognized by pushdown automata (PDA), it suffices to determine the state transition rules and stack action rules of the corresponding PDA. An approach would be to train a recurrent neural network (RNN) to classify the sample data and attempt to extract these PDA rules. But neural networks are not a priori aware of the structure of a PDA and would likely require many samples to infer this structure. Furthermore, extracting the PDA rules from the RNN is nontrivial. We build a RNN specifically structured like a PDA, where weights correspond directly to the PDA rules. This requires a stack architecture that is somehow differentiable (to enable gradient-based learning) and stable (an unstable stack will show deteriorating performance with longer strings). We propose a stack architecture that is differentiable and that provably exhibits orbital stability. Using this stack, we construct a neural network that provably approximates a PDA for strings of arbitrary length. Moreover, our model and method of proof can easily be generalized to other state machines, such as a Turing Machine.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/09/2020

Learning Context-Free Languages with Nondeterministic Stack RNNs

We present a differentiable stack data structure that simultaneously and...
research
11/15/2017

The Neural Network Pushdown Automaton: Model, Stack and Learning Simulations

In order for neural networks to learn complex languages or grammars, the...
research
09/05/2021

Learning Hierarchical Structures with Differentiable Nondeterministic Stacks

Learning hierarchical structures in sequential data – from simple algori...
research
02/10/2021

Differentiable Generative Phonology

The goal of generative phonology, as formulated by Chomsky and Halle (19...
research
09/07/2019

The Neural State Pushdown Automata

In order to learn complex grammars, recurrent neural networks (RNNs) req...
research
05/15/2019

Techniques for Inferring Context-Free Lindenmayer Systems With Genetic Algorithm

Lindenmayer systems (L-systems) are a formal grammar system, where the m...
research
06/04/2019

Finding Syntactic Representations in Neural Stacks

Neural network architectures have been augmented with differentiable sta...

Please sign up or login with your details

Forgot password? Click here to reset