This looks more like that: Enhancing Self-Explaining Models by Prototypical Relevance Propagation

08/27/2021
by   Srishti Gautam, et al.
0

Current machine learning models have shown high efficiency in solving a wide variety of real-world problems. However, their black box character poses a major challenge for the understanding and traceability of the underlying decision-making strategies. As a remedy, many post-hoc explanation and self-explanatory methods have been developed to interpret the models' behavior. These methods, in addition, enable the identification of artifacts that can be learned by the model as class-relevant features. In this work, we provide a detailed case study of the self-explaining network, ProtoPNet, in the presence of a spectrum of artifacts. Accordingly, we identify the main drawbacks of ProtoPNet, especially, its coarse and spatially imprecise explanations. We address these limitations by introducing Prototypical Relevance Propagation (PRP), a novel method for generating more precise model-aware explanations. Furthermore, in order to obtain a clean dataset, we propose to use multi-view clustering strategies for segregating the artifact images using the PRP explanations, thereby suppressing the potential artifact learning in the models.

READ FULL TEXT

page 1

page 4

page 5

page 6

page 9

page 10

page 11

page 13

research
09/12/2020

MeLIME: Meaningful Local Explanation for Machine Learning Models

Most state-of-the-art machine learning algorithms induce black-box model...
research
02/15/2023

Streamlining models with explanations in the learning loop

Several explainable AI methods allow a Machine Learning user to get insi...
research
10/09/2021

Self-explaining Neural Network with Plausible Explanations

Explaining the predictions of complex deep learning models, often referr...
research
02/22/2021

Explainers in the Wild: Making Surrogate Explainers Robust to Distortions through Perception

Explaining the decisions of models is becoming pervasive in the image pr...
research
12/19/2022

Explaining Classifications to Non Experts: An XAI User Study of Post Hoc Explanations for a Classifier When People Lack Expertise

Very few eXplainable AI (XAI) studies consider how users understanding o...
research
11/10/2020

Debugging Tests for Model Explanations

We investigate whether post-hoc model explanations are effective for dia...

Please sign up or login with your details

Forgot password? Click here to reset