Interpretations are useful: penalizing explanations to align neural networks with prior knowledge

09/30/2019
by   Laura Rieger, et al.
20

For an explanation of a deep learning model to be effective, it must provide both insight into a model and suggest a corresponding action in order to achieve some objective. Too often, the litany of proposed explainable deep learning methods stop at the first step, providing practitioners with insight into a model, but no way to act on it. In this paper, we propose contextual decomposition explanation penalization (CDEP), a method which enables practitioners to leverage existing explanation methods in order to increase the predictive accuracy of deep learning models. In particular, when shown that a model has incorrectly assigned importance to some features, CDEP enables practitioners to correct these errors by directly regularizing the provided explanations. Using explanations provided by contextual decomposition (CD) (Murdoch et al., 2018), we demonstrate the ability of our method to increase performance on an array of toy and real datasets.

READ FULL TEXT

page 5

page 6

page 14

page 15

page 16

page 17

research
02/15/2022

Contextual Importance and Utility: aTheoretical Foundation

This paper provides new theory to support to the eXplainable AI (XAI) me...
research
04/29/2021

Correcting Classification: A Bayesian Framework Using Explanation Feedback to Improve Classification Abilities

Neural networks (NNs) have shown high predictive performance, however, w...
research
04/29/2021

A First Look: Towards Explainable TextVQA Models via Visual and Textual Explanations

Explainable deep learning models are advantageous in many situations. Pr...
research
12/11/2020

Dependency Decomposition and a Reject Option for Explainable Models

Deploying machine learning models in safety-related do-mains (e.g. auton...
research
06/25/2020

Explainable CNN-attention Networks (C-Attention Network) for Automated Detection of Alzheimer's Disease

In this work, we propose three explainable deep learning architectures t...
research
11/08/2019

Towards Hierarchical Importance Attribution: Explaining Compositional Semantics for Neural Sequence Models

The impressive performance of neural networks on natural language proces...
research
03/25/2023

Learning with Explanation Constraints

While supervised learning assumes the presence of labeled data, we may h...

Please sign up or login with your details

Forgot password? Click here to reset