Good-looking but Lacking Faithfulness: Understanding Local Explanation Methods through Trend-based Testing

09/09/2023
by   Jinwen He, et al.
0

While enjoying the great achievements brought by deep learning (DL), people are also worried about the decision made by DL models, since the high degree of non-linearity of DL models makes the decision extremely difficult to understand. Consequently, attacks such as adversarial attacks are easy to carry out, but difficult to detect and explain, which has led to a boom in the research on local explanation methods for explaining model decisions. In this paper, we evaluate the faithfulness of explanation methods and find that traditional tests on faithfulness encounter the random dominance problem, , the random selection performs the best, especially for complex data. To further solve this problem, we propose three trend-based faithfulness tests and empirically demonstrate that the new trend tests can better assess faithfulness than traditional tests on image, natural language and security tasks. We implement the assessment system and evaluate ten popular explanation methods. Benefiting from the trend tests, we successfully assess the explanation methods on complex data for the first time, bringing unprecedented discoveries and inspiring future research. Downstream tasks also greatly benefit from the tests. For example, model debugging equipped with faithful explanation methods performs much better for detecting and correcting accuracy and security problems.

READ FULL TEXT

page 13

page 19

research
04/29/2021

Twin Systems for DeepCBR: A Menagerie of Deep Learning and Case-Based Reasoning Pairings for Explanation and Data Augmentation

Recently, it has been proposed that fruitful synergies may exist between...
research
03/30/2022

Example-based Explanations with Adversarial Attacks for Respiratory Sound Analysis

Respiratory sound classification is an important tool for remote screeni...
research
10/09/2019

Explaining Deep Learning-Based Networked Systems

While deep learning (DL)-based networked systems have shown great potent...
research
12/14/2016

Attentive Explanations: Justifying Decisions and Pointing to the Evidence

Deep models are the defacto standard in visual decision models due to th...
research
04/05/2023

Rethinking the Trigger-injecting Position in Graph Backdoor Attack

Backdoor attacks have been demonstrated as a security threat for machine...
research
10/15/2022

Providing Error Detection for Deep Learning Image Classifiers Using Self-Explainability

This paper proposes a self-explainable Deep Learning (SE-DL) system for ...

Please sign up or login with your details

Forgot password? Click here to reset