A Diagnostic Study of Explainability Techniques for Text Classification

09/25/2020
by   Pepa Atanasova, et al.
0

Recent developments in machine learning have introduced models that approach human performance at the cost of increased architectural complexity. Efforts to make the rationales behind the models' predictions transparent have inspired an abundance of new explainability techniques. Provided with an already trained model, they compute saliency scores for the words of an input instance. However, there exists no definitive guide on (i) how to choose such a technique given a particular application task and model architecture, and (ii) the benefits and drawbacks of using each such technique. In this paper, we develop a comprehensive list of diagnostic properties for evaluating existing explainability techniques. We then employ the proposed list to compare a set of diverse explainability techniques on downstream text classification tasks and neural network architectures. We also compare the saliency scores assigned by the explainability techniques with human annotations of salient input regions to find relations between a model's performance and the agreement of its rationales with human ones. Overall, we find that the gradient-based explanations perform best across tasks and model architectures, and we present further insights into the properties of the reviewed explainability techniques.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/10/2022

Towards Human-Centred Explainability Benchmarks For Text Classification

Progress on many Natural Language Processing (NLP) tasks, such as text c...
research
09/13/2019

Explainable Machine Learning in Deployment

Explainable machine learning seeks to provide various stakeholders with ...
research
04/05/2021

Explainability-aided Domain Generalization for Image Classification

Traditionally, for most machine learning settings, gaining some degree o...
research
07/17/2022

Towards Explainability in NLP: Analyzing and Calculating Word Saliency through Word Properties

The wide use of black-box models in natural language processing brings g...
research
08/12/2022

The Weighting Game: Evaluating Quality of Explainability Methods

The objective of this paper is to assess the quality of explanation heat...
research
05/31/2019

Explainability Techniques for Graph Convolutional Networks

Graph Networks are used to make decisions in potentially complex scenari...
research
02/13/2023

Gradient-Based Automated Iterative Recovery for Parameter-Efficient Tuning

Pretrained large language models (LLMs) are able to solve a wide variety...

Please sign up or login with your details

Forgot password? Click here to reset