Learning to Recombine and Resample Data for Compositional Generalization

10/08/2020
by   Ekin Akyürek, et al.
0

Flexible neural models outperform grammar- and automaton-based counterparts on a variety of sequence modeling tasks. However, neural models perform poorly in settings requiring compositional generalization beyond the training data – particularly to rare or unseen subsequences. Past work has found symbolic scaffolding (e.g. grammars or automata) essential in these settings. Here we present a family of learned data augmentation schemes that support a large category of compositional generalizations without appeal to latent symbolic structure. Our approach to data augmentation has two components: recombination of original training examples via a prototype-based generative model and resampling of generated examples to encourage extrapolation. Training an ordinary neural sequence model on a dataset augmented with recombined and resampled examples significantly improves generalization in two language processing problems—instruction following (SCAN) and morphological analysis (Sigmorphon 2018)—where our approach enables learning of new constructions and tenses from as few as eight initial examples.

READ FULL TEXT
research
12/14/2021

Improving Compositional Generalization with Latent Structure and Data Augmentation

Generic unstructured neural networks have been shown to struggle on out-...
research
11/18/2020

Sequence-Level Mixed Sample Data Augmentation

Despite their empirical success, neural networks still have difficulty c...
research
04/21/2019

Good-Enough Compositional Data Augmentation

We propose a simple data augmentation protocol aimed at providing a comp...
research
08/15/2020

Compositional Generalization via Neural-Symbolic Stack Machines

Despite achieving tremendous success, existing deep learning models have...
research
04/05/2021

Paired Examples as Indirect Supervision in Latent Decision Models

Compositional, structured models are appealing because they explicitly d...
research
01/30/2022

Compositionality as Lexical Symmetry

Standard deep network models lack the inductive biases needed to general...
research
04/30/2020

When does data augmentation help generalization in NLP?

Neural models often exploit superficial ("weak") features to achieve goo...

Please sign up or login with your details

Forgot password? Click here to reset