Who Explains the Explanation? Quantitatively Assessing Feature Attribution Methods

09/28/2021
by   Anna Arias-Duart, et al.
1

AI explainability seeks to increase the transparency of models, making them more trustworthy in the process. The need for transparency has been recently motivated by the emergence of deep learning models, which are particularly obscure by nature. Even in the domain of images, where deep learning has succeeded the most, explainability is still poorly assessed. Multiple feature attribution methods have been proposed in the literature with the purpose of explaining a DL model's behavior using visual queues, but no standardized metrics to assess or select these methods exist. In this paper we propose a novel evaluation metric – the Focus – designed to quantify the faithfulness of explanations provided by feature attribution methods, such as LRP or GradCAM. First, we show the robustness of the metric through randomization experiments, and then use Focus to evaluate and compare three popular explainability techniques using multiple architectures and datasets. Our results find LRP and GradCAM to be consistent and reliable, the former being more accurate for high performing models, while the latter remains most competitive even when applied to poorly performing models. Finally, we identify a strong relation between Focus and factors like model architecture and task, unveiling a new unsupervised approach for the assessment of models.

READ FULL TEXT
research
09/26/2020

Quantitative and Qualitative Evaluation of Explainable Deep Learning Methods for Ophthalmic Diagnosis

Background: The lack of explanations for the decisions made by algorithm...
research
11/23/2022

Evaluating Feature Attribution Methods for Electrocardiogram

The performance of cardiac arrhythmia detection with electrocardiograms(...
research
07/04/2022

A Framework for Auditing Multilevel Models using Explainability Methods

Applications of multilevel models usually result in binary classificatio...
research
03/29/2021

Efficient Explanations from Empirical Explainers

Amid a discussion about Green AI in which we see explainability neglecte...
research
12/06/2021

What I Cannot Predict, I Do Not Understand: A Human-Centered Evaluation Framework for Explainability Methods

A multitude of explainability methods and theoretical evaluation scores ...
research
11/06/2021

"How Does It Detect A Malicious App?" Explaining the Predictions of AI-based Android Malware Detector

AI methods have been proven to yield impressive performance on Android m...
research
01/17/2023

Negative Flux Aggregation to Estimate Feature Attributions

There are increasing demands for understanding deep neural networks' (DN...

Please sign up or login with your details

Forgot password? Click here to reset