Benchmarking Neural Network Generalization for Grammar Induction

08/16/2023
by   Nur Lan, et al.
0

How well do neural networks generalize? Even for grammar induction tasks, where the target generalization is fully known, previous works have left the question open, testing very limited ranges beyond the training set and using different success criteria. We provide a measure of neural network generalization based on fully specified formal languages. Given a model and a formal grammar, the method assigns a generalization score representing how well a model generalizes to unseen samples in inverse relation to the amount of data it was trained on. The benchmark includes languages such as a^nb^n, a^nb^nc^n, a^nb^mc^n+m, and Dyck-1 and 2. We evaluate selected architectures using the benchmark and find that networks trained with a Minimum Description Length objective (MDL) generalize better and using less data than networks trained using standard loss functions. The benchmark is available at https://github.com/taucompling/bliss.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/31/2021

Minimum Description Length Recurrent Neural Networks

We train neural networks to optimize a Minimum Description Length score,...
research
07/29/2020

The Return of Lexical Dependencies: Neural Lexicalized PCFGs

In this paper we demonstrate that context free grammar (CFG) based metho...
research
10/28/2017

Inducing Regular Grammars Using Recurrent Neural Networks

Grammar induction is the task of learning a grammar from a set of exampl...
research
02/26/2020

Predicting Neural Network Accuracy from Weights

We study the prediction of the accuracy of a neural network given only i...
research
04/11/2019

Frequency vs. Association for Constraint Selection in Usage-Based Construction Grammar

A usage-based Construction Grammar (CxG) posits that slot-constraints ge...
research
03/23/2018

Generalization Challenges for Neural Architectures in Audio Source Separation

Recent work has shown that recurrent neural networks can be trained to s...
research
05/31/2019

Constructive Type-Logical Supertagging with Self-Attention Networks

We propose a novel application of self-attention networks towards gramma...

Please sign up or login with your details

Forgot password? Click here to reset