DeepAI
Log In Sign Up

Explaining the Road Not Taken

03/27/2021
by   Hua Shen, et al.
0

It is unclear if existing interpretations of deep neural network models respond effectively to the needs of users. This paper summarizes the common forms of explanations (such as feature attribution, decision rules, or probes) used in over 200 recent papers about natural language processing (NLP), and compares them against user questions collected in the XAI Question Bank. We found that although users are interested in explanations for the road not taken – namely, why the model chose one result and not a well-defined, seemly similar legitimate counterpart – most model interpretations cannot answer these questions.

READ FULL TEXT
08/13/2018

Learning Explanations from Language Data

PatternAttribution is a recent method, introduced in the vision domain, ...
06/01/2020

Aligning Faithful Interpretations with their Social Attribution

We find that the requirement of model interpretations to be faithful is ...
05/23/2022

A Question-Answer Driven Approach to Reveal Affirmative Interpretations from Verbal Negations

This paper explores a question-answer driven approach to reveal affirmat...
04/21/2019

NeuronBlocks -- Building Your NLP DNN Models Like Playing Lego

When building deep neural network models for natural language processing...
04/09/2021

Explaining Neural Network Predictions on Sentence Pairs via Learning Word-Group Masks

Explaining neural network models is important for increasing their trust...
04/01/2022

Extracting Rules from Neural Networks with Partial Interpretations

We investigate the problem of extracting rules, expressed in Horn logic,...

Code Repositories

human-centered-exnlp.github.io

None


view repo