On Explaining Your Explanations of BERT: An Empirical Study with Sequence Classification

01/01/2021
by   Zhengxuan Wu, et al.
0

BERT, as one of the pretrianed language models, attracts the most attention in recent years for creating new benchmarks across GLUE tasks via fine-tuning. One pressing issue is to open up the blackbox and explain the decision makings of BERT. A number of attribution techniques have been proposed to explain BERT models, but are often limited to sequence to sequence tasks. In this paper, we adapt existing attribution methods on explaining decision makings of BERT in sequence classification tasks. We conduct extensive analyses of four existing attribution methods by applying them to four different datasets in sentiment analysis. We compare the reliability and robustness of each method via various ablation studies. Furthermore, we test whether attribution methods explain generalized semantics across semantically similar tasks. Our work provides solid guidance for using attribution methods to explain decision makings of BERT for downstream classification tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
08/31/2021

Explaining Classes through Word Attribution

In recent years, several methods have been proposed for explaining indiv...
research
10/22/2021

Double Trouble: How to not explain a text classifier's decisions using counterfactuals synthesized by masked language models?

Explaining how important each input feature is to a classifier's decisio...
research
12/28/2020

Enhanced Regularizers for Attributional Robustness

Deep neural networks are the default choice of learning models for compu...
research
05/24/2023

Scale Matters: Attribution Meets the Wavelet Domain to Explain Model Sensitivity to Image Corruptions

Neural networks have shown remarkable performance in computer vision, bu...
research
03/23/2022

An Empirical Study of Memorization in NLP

A recent study by Feldman (2020) proposed a long-tail theory to explain ...
research
08/04/2023

Explaining Relation Classification Models with Semantic Extents

In recent years, the development of large pretrained language models, su...
research
11/22/2022

Explaining YOLO: Leveraging Grad-CAM to Explain Object Detections

We investigate the problem of explainability for visual object detectors...

Please sign up or login with your details

Forgot password? Click here to reset