Generating Hierarchical Explanations on Text Classification Without Connecting Rules

10/24/2022
by   Yiming Ju, et al.
0

The opaqueness of deep NLP models has motivated the development of methods for interpreting how deep models predict. Recently, work has introduced hierarchical attribution, which produces a hierarchical clustering of words, along with an attribution score for each cluster. However, existing work on hierarchical attribution all follows the connecting rule, limiting the cluster to a continuous span in the input text. We argue that the connecting rule as an additional prior may undermine the ability to reflect the model decision process faithfully. To this end, we propose to generate hierarchical explanations without the connecting rule and introduce a framework for generating hierarchical clusters. Experimental results and further analysis show the effectiveness of the proposed method in providing high-quality explanations for reflecting model predicting process.

READ FULL TEXT

page 7

page 8

research
04/04/2020

Generating Hierarchical Explanations on Text Classification via Feature Interaction Detection

Generating explanations for neural networks has become crucial for their...
research
02/14/2022

Measurably Stronger Explanation Reliability via Model Canonization

While rule-based attribution methods have proven useful for providing lo...
research
12/18/2022

Estimating the Adversarial Robustness of Attributions in Text with Transformers

Explanations are crucial parts of deep neural network (DNN) classifiers....
research
03/20/2021

Boundary Attributions Provide Normal (Vector) Explanations

Recent work on explaining Deep Neural Networks (DNNs) focuses on attribu...
research
11/08/2018

Looking Deeper into Deep Learning Model: Attribution-based Explanations of TextCNN

Layer-wise Relevance Propagation (LRP) and saliency maps have been recen...
research
03/29/2021

Efficient Explanations from Empirical Explainers

Amid a discussion about Green AI in which we see explainability neglecte...
research
07/15/2020

Learning Invariances for Interpretability using Supervised VAE

We propose to learn model invariances as a means of interpreting a model...

Please sign up or login with your details

Forgot password? Click here to reset