EvalAttAI: A Holistic Approach to Evaluating Attribution Maps in Robust and Non-Robust Models

03/15/2023
by   Ian E. Nielsen, et al.
0

The expansion of explainable artificial intelligence as a field of research has generated numerous methods of visualizing and understanding the black box of a machine learning model. Attribution maps are generally used to highlight the parts of the input image that influence the model to make a specific decision. On the other hand, the robustness of machine learning models to natural noise and adversarial attacks is also being actively explored. This paper focuses on evaluating methods of attribution mapping to find whether robust neural networks are more explainable. We explore this problem within the application of classification for medical imaging. Explainability research is at an impasse. There are many methods of attribution mapping, but no current consensus on how to evaluate them and determine the ones that are the best. Our experiments on multiple datasets (natural and medical imaging) and various attribution methods reveal that two popular evaluation metrics, Deletion and Insertion, have inherent limitations and yield contradictory results. We propose a new explainability faithfulness metric (called EvalAttAI) that addresses the limitations of prior metrics. Using our novel evaluation, we found that Bayesian deep neural networks using the Variational Density Propagation technique were consistently more explainable when used with the best performing attribution method, the Vanilla Gradient. However, in general, various types of robust neural networks may not be more explainable, despite these models producing more visually plausible attribution maps.

READ FULL TEXT

page 1

page 3

page 4

page 6

page 9

page 12

research
07/20/2023

Is Grad-CAM Explainable in Medical Images?

Explainable Deep Learning has gained significant attention in the field ...
research
02/22/2022

Evaluating Feature Attribution Methods in the Image Domain

Feature attribution maps are a popular approach to highlight the most im...
research
08/23/2021

Longitudinal Distance: Towards Accountable Instance Attribution

Previous research in interpretable machine learning (IML) and explainabl...
research
05/25/2022

How explainable are adversarially-robust CNNs?

Three important criteria of existing convolutional neural networks (CNNs...
research
06/15/2023

Improving Explainability of Disentangled Representations using Multipath-Attribution Mappings

Explainable AI aims to render model behavior understandable by humans, w...
research
08/26/2020

Making Neural Networks Interpretable with Attribution: Application to Implicit Signals Prediction

Explaining recommendations enables users to understand whether recommend...
research
03/21/2023

Better Understanding Differences in Attribution Methods via Systematic Evaluations

Deep neural networks are very successful on many vision tasks, but hard ...

Please sign up or login with your details

Forgot password? Click here to reset