Understanding Attention in Machine Reading Comprehension

08/26/2021
by   Yiming Cui, et al.
0

Achieving human-level performance on some of Machine Reading Comprehension (MRC) datasets is no longer challenging with the help of powerful Pre-trained Language Models (PLMs). However, the internal mechanism of these artifacts still remains unclear, placing an obstacle for further understanding these models. This paper focuses on conducting a series of analytical experiments to examine the relations between the multi-head self-attention and the final performance, trying to analyze the potential explainability in PLM-based MRC models. We perform quantitative analyses on SQuAD (English) and CMRC 2018 (Chinese), two span-extraction MRC datasets, on top of BERT, ALBERT, and ELECTRA in various aspects. We discover that passage-to-question and passage understanding attentions are the most important ones, showing strong correlations to the final performance than other parts. Through visualizations and case studies, we also observe several general findings on the attention maps, which could be helpful to understand how these models solve the questions.

READ FULL TEXT

page 4

page 6

research
11/13/2020

Unsupervised Explanation Generation for Machine Reading Comprehension

With the blooming of various Pre-trained Language Models (PLMs), Machine...
research
05/10/2021

ExpMRC: Explainability Evaluation for Machine Reading Comprehension

Achieving human-level performance on some of Machine Reading Comprehensi...
research
05/31/2021

A Multilingual Modeling Method for Span-Extraction Reading Comprehension

Span-extraction reading comprehension models have made tremendous advanc...
research
05/24/2021

Using Adversarial Attacks to Reveal the Statistical Bias in Machine Reading Comprehension Models

Pre-trained language models have achieved human-level performance on man...
research
08/28/2019

Discourse-Aware Semantic Self-Attention for Narrative Reading Comprehension

In this work, we propose to use linguistic annotations as a basis for a ...
research
07/13/2021

Deep Neural Networks Evolve Human-like Attention Distribution during Reading Comprehension

Attention is a key mechanism for information selection in both biologica...
research
08/20/2022

Trigger-free Event Detection via Derangement Reading Comprehension

Event detection (ED), aiming to detect events from texts and categorize ...

Please sign up or login with your details

Forgot password? Click here to reset