Attention Flows are Shapley Value Explanations

05/31/2021
by   Kawin Ethayarajh, et al.
0

Shapley Values, a solution to the credit assignment problem in cooperative game theory, are a popular type of explanation in machine learning, having been used to explain the importance of features, embeddings, and even neurons. In NLP, however, leave-one-out and attention-based explanations still predominate. Can we draw a connection between these different methods? We formally prove that – save for the degenerate case – attention weights and leave-one-out values cannot be Shapley Values. Attention flow is a post-processed variant of attention weights obtained by running the max-flow algorithm on the attention graph. Perhaps surprisingly, we prove that attention flows are indeed Shapley Values, at least at the layerwise level. Given the many desirable theoretical qualities of Shapley Values – which has driven their adoption among the ML community – we argue that NLP practitioners should, when possible, adopt attention flow explanations alongside more traditional ones.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/25/2020

Problems with Shapley-value-based explanations as feature importance measures

Game-theoretic formulations of feature importance have become popular as...
research
01/26/2022

Attention cannot be an Explanation

Attention based explanations (viz. saliency maps), by providing interpre...
research
02/26/2019

Attention is not Explanation

Attention mechanisms have seen wide adoption in neural NLP models. In ad...
research
05/02/2020

Quantifying Attention Flow in Transformers

In the Transformer model, "self-attention" combines information from att...
research
09/17/2019

The Explanation Game: Explaining Machine Learning Models with Cooperative Game Theory

Recently, a number of techniques have been proposed to explain a machine...
research
08/09/2021

Improved Feature Importance Computations for Tree Models: Shapley vs. Banzhaf

Shapley values are one of the main tools used to explain predictions of ...
research
10/27/2020

Shapley Flow: A Graph-based Approach to Interpreting Model Predictions

Many existing approaches for estimating feature importance are problemat...

Please sign up or login with your details

Forgot password? Click here to reset