On the Transferability of Minimal Prediction Preserving Inputs in Question Answering

09/17/2020
by   Shayne Longpre, et al.
0

Recent work (Feng et al., 2018) establishes the presence of short, uninterpretable input fragments that yield high confidence and accuracy in neural models. We refer to these as Minimal Prediction Preserving Inputs (MPPIs). In the context of question answering, we investigate competing hypotheses for the existence of MPPIs, including poor posterior calibration of neural models, lack of pretraining, and "dataset bias" (where a model learns to attend to spurious, non-generalizable cues in the training data). We discover a perplexing invariance of MPPIs to random training seed, model architecture, pretraining, and training domain. MPPIs demonstrate remarkable transferability across domains - closing half the gap between models' performance on comparably short queries and original queries. Additionally, penalizing over-confidence on MPPIs fails to improve either generalization or adversarial robustness. These results suggest the interpretability of MPPIs is insufficient to characterize generalization capacity of these models. We hope this focused investigation encourages a more systematic analysis of model behavior outside of the human interpretable distribution of examples.

READ FULL TEXT
research
05/21/2018

Efficient and Robust Question Answering from Minimal Context over Documents

Neural models for question answering (QA) over documents have achieved s...
research
06/18/2018

Comparative Analysis of Neural QA models on SQuAD

The task of Question Answering has gained prominence in the past few dec...
research
07/01/2020

Latent Compositional Representations Improve Systematic Generalization in Grounded Question Answering

Answering questions that involve multi-step reasoning requires decomposi...
research
04/16/2020

Bridging Anaphora Resolution as Question Answering

Most previous studies on bridging anaphora resolution (Poesio et al., 20...
research
03/20/2022

Calibration of Machine Reading Systems at Scale

In typical machine learning systems, an estimate of the probability of t...
research
04/20/2018

Right Answer for the Wrong Reason: Discovery and Mitigation

Exposing the weaknesses of neural models is crucial for improving their ...
research
12/20/2022

Socratic Pretraining: Question-Driven Pretraining for Controllable Summarization

In long document controllable summarization, where labeled data is scarc...

Please sign up or login with your details

Forgot password? Click here to reset