Selective Question Answering under Domain Shift

06/16/2020
by   Amita Kamath, et al.
0

To avoid giving wrong answers, question answering (QA) models need to know when to abstain from answering. Moreover, users often ask questions that diverge from the model's training data, making errors more likely and thus abstention more critical. In this work, we propose the setting of selective question answering under domain shift, in which a QA model is tested on a mixture of in-domain and out-of-domain data, and must answer (i.e., not abstain on) as many questions as possible while maintaining high accuracy. Abstention policies based solely on the model's softmax probabilities fare poorly, since models are overconfident on out-of-domain inputs. Instead, we train a calibrator to identify inputs on which the QA model errs, and abstain when it predicts an error is likely. Crucially, the calibrator benefits from observing the model's behavior on out-of-domain data, even if from a different domain than the test data. We combine this method with a SQuAD-trained QA model and evaluate on mixtures of SQuAD and five other QA datasets. Our method answers 56 the model's probabilities only answers 48

READ FULL TEXT
research
04/18/2021

Can NLI Models Verify QA Systems' Predictions?

To build robust question answering systems, we need the ability to verif...
research
08/20/2017

Learning to Paraphrase for Question Answering

Question answering (QA) systems are sensitive to the many different ways...
research
04/09/2022

Extending the Scope of Out-of-Domain: Examining QA models in multiple subdomains

Past works that investigate out-of-domain performance of QA systems have...
research
12/19/2022

Tokenization Consistency Matters for Generative Models on Extractive NLP Tasks

Generative models have been widely applied to solve extractive tasks, wh...
research
11/29/2022

Penalizing Confident Predictions on Largely Perturbed Inputs Does Not Improve Out-of-Distribution Generalization in Question Answering

Question answering (QA) models are shown to be insensitive to large pert...
research
10/02/2017

Building Chatbots from Forum Data: Model Selection Using Question Answering Metrics

We propose to use question answering (QA) data from Web forums to train ...
research
02/13/2021

PAQ: 65 Million Probably-Asked Questions and What You Can Do With Them

Open-domain Question Answering models which directly leverage question-a...

Please sign up or login with your details

Forgot password? Click here to reset