Evaluating Explanations for Reading Comprehension with Realistic Counterfactuals

04/09/2021
by   Xi Ye, et al.
0

Token-level attributions have been extensively studied to explain model predictions for a wide range of classification tasks in NLP (e.g., sentiment analysis), but such explanation techniques are less explored for machine reading comprehension (RC) tasks. Although the transformer-based models used here are identical to those used for classification, the underlying reasoning these models perform is very different and different types of explanations are required. We propose a methodology to evaluate explanations: an explanation should allow us to understand the RC model's high-level behavior with respect to a set of realistic counterfactual input scenarios. We define these counterfactuals for several RC settings, and by connecting explanation techniques' outputs to high-level model behavior, we can evaluate how useful different explanations really are. Our analysis suggests that pairwise explanation techniques are better suited to RC than token-level attributions, which are often unfaithful in the scenarios we consider. We additionally propose an improvement to an attention-based attribution technique, resulting in explanations which better reveal the model's behavior.

READ FULL TEXT

page 16

page 17

research
11/13/2020

Unsupervised Explanation Generation for Machine Reading Comprehension

With the blooming of various Pre-trained Language Models (PLMs), Machine...
research
03/02/2021

Contrastive Explanations for Model Interpretability

Contrastive explanations clarify why an event occurred in contrast to an...
research
09/15/2022

Machine Reading, Fast and Slow: When Do Models "Understand" Language?

Two of the most fundamental challenges in Natural Language Understanding...
research
05/02/2020

Teaching Machine Comprehension with Compositional Explanations

Advances in extractive machine reading comprehension (MRC) rely heavily ...
research
09/14/2021

Summarize-then-Answer: Generating Concise Explanations for Multi-hop Reading Comprehension

How can we generate concise explanations for multi-hop Reading Comprehen...
research
10/17/2022

On the Impact of Temporal Concept Drift on Model Explanations

Explanation faithfulness of model predictions in natural language proces...
research
05/27/2022

CEBaB: Estimating the Causal Effects of Real-World Concepts on NLP Model Behavior

The increasing size and complexity of modern ML systems has improved the...

Please sign up or login with your details

Forgot password? Click here to reset