Symbolic Brittleness in Sequence Models: on Systematic Generalization in Symbolic Mathematics

09/28/2021
by   Sean Welleck, et al.
0

Neural sequence models trained with maximum likelihood estimation have led to breakthroughs in many tasks, where success is defined by the gap between training and test performance. However, their ability to achieve stronger forms of generalization remains unclear. We consider the problem of symbolic mathematical integration, as it requires generalizing systematically beyond the test set. We develop a methodology for evaluating generalization that takes advantage of the problem domain's structure and access to a verifier. Despite promising in-distribution performance of sequence-to-sequence models in this domain, we demonstrate challenges in achieving robustness, compositionality, and out-of-distribution generalization, through both carefully constructed manual test suites and a genetic algorithm that automatically finds large collections of failures in a controllable manner. Our investigation highlights the difficulty of generalizing well with the predominant modeling and learning approach, and the importance of evaluating beyond the test set, across different aspects of generalization.

READ FULL TEXT
research
05/03/2018

The Fine Line between Linguistic Generalization and Failure in Seq2Seq-Attention Models

Seq2Seq based neural architectures have become the go-to architecture to...
research
12/02/2019

Deep Learning for Symbolic Mathematics

Neural networks have a reputation for being better at solving statistica...
research
10/07/2021

Pretrained Language Models are Symbolic Mathematics Solvers too!

Solving symbolic mathematics has always been of in the arena of human in...
research
10/30/2017

Sequence-to-Sequence ASR Optimization via Reinforcement Learning

Despite the success of sequence-to-sequence approaches in automatic spee...
research
08/15/2021

Exploring Generalization Ability of Pretrained Language Models on Arithmetic and Logical Reasoning

To quantitatively and intuitively explore the generalization ability of ...
research
04/23/2021

Learning to Learn to be Right for the Right Reasons

Improving model generalization on held-out data is one of the core objec...
research
02/19/2022

Do Transformers use variable binding?

Increasing the explainability of deep neural networks (DNNs) requires ev...

Please sign up or login with your details

Forgot password? Click here to reset