Aggregating explainability methods for neural networks stabilizes explanations

03/01/2019
by   Laura Rieger, et al.
0

Despite a growing literature on explaining neural networks, no consensus has been reached on how to explain a neural network decision or how to evaluate an explanation. In fact, most works rely on manually assessing the explanation to evaluate the quality of a method. This injects uncertainty in the explanation process along several dimensions: Which explanation method to apply? Who should we ask to evaluate it and which criteria should be used for the evaluation? Our contributions in this paper are twofold. First, we investigate schemes to combine explanation methods and reduce model uncertainty to obtain a single aggregated explanation. Our findings show that the aggregation is more robust, well-aligned with human explanations and can attribute relevance to a broader set of features (completeness). Second, we propose a novel way of evaluating explanation methods that circumvents the need for manual evaluation and is not reliant on the alignment of neural networks and humans decision processes.

READ FULL TEXT

page 2

page 4

page 7

page 8

research
03/04/2022

Do Explanations Explain? Model Knows Best

It is a mystery which input features contribute to a neural network's ou...
research
08/12/2022

The Weighting Game: Evaluating Quality of Explainability Methods

The objective of this paper is to assess the quality of explanation heat...
research
04/09/2019

Software and application patterns for explanation methods

Deep neural networks successfully pervaded many applications domains and...
research
07/04/2022

Fidelity of Ensemble Aggregation for Saliency Map Explanations using Bayesian Optimization Techniques

In recent years, an abundance of feature attribution methods for explain...
research
01/19/2018

Evaluating neural network explanation methods using hybrid documents and morphological prediction

We propose two novel paradigms for evaluating neural network explanation...
research
03/31/2020

Explaining Motion Relevance for Activity Recognition in Video Deep Learning Models

A small subset of explainability techniques developed initially for imag...
research
09/13/2021

Explaining Deep Learning Representations by Tracing the Training Process

We propose a novel explanation method that explains the decisions of a d...

Please sign up or login with your details

Forgot password? Click here to reset