Towards Prediction Explainability through Sparse Communication

04/28/2020
by   Marcos V. Treviso, et al.
0

Explainability is a topic of growing importance in NLP. In this work, we provide a unified perspective of explainability as a communication problem between an explainer and a layperson about a classifier's decision. We use this framework to compare several prior approaches for extracting explanations, including gradient methods, representation erasure, and attention mechanisms, in terms of their communication success. In addition, we reinterpret these methods at the light of classical feature selection, and we use this as inspiration to propose new embedded methods for explainability, through the use of selective, sparse attention. Experiments in text classification, natural language entailment, and machine translation, using different configurations of explainers and laypeople (including both machines and humans), reveal an advantage of attention-based explainers over gradient and erasure methods. Furthermore, human evaluation experiments show promising results with post-hoc explainers trained to optimize communication success and faithfulness.

READ FULL TEXT
research
01/28/2022

Rethinking Attention-Model Explainability through Faithfulness Violation Test

Attention mechanisms are dominating the explainability of deep models. T...
research
02/28/2023

Multi-Layer Attention-Based Explainability via Transformers for Tabular Data

We propose a graph-oriented attention-based explainability method for ta...
research
11/10/2022

Towards Human-Centred Explainability Benchmarks For Text Classification

Progress on many Natural Language Processing (NLP) tasks, such as text c...
research
08/08/2023

Semantic Interpretation and Validation of Graph Attention-based Explanations for GNN Models

In this work, we propose a methodology for investigating the application...
research
07/12/2021

Quantifying Explainability in NLP and Analyzing Algorithms for Performance-Explainability Tradeoff

The healthcare domain is one of the most exciting application areas for ...
research
04/12/2022

Maximum Entropy Baseline for Integrated Gradients

Integrated Gradients (IG), one of the most popular explainability method...
research
07/27/2023

Explainable Techniques for Analyzing Flow Cytometry Cell Transformers

Explainability for Deep Learning Models is especially important for clin...

Please sign up or login with your details

Forgot password? Click here to reset