Local Explanation Methods for Deep Neural Networks Lack Sensitivity to Parameter Values

10/08/2018
by   Julius Adebayo, et al.
2

Explaining the output of a complicated machine learning model like a deep neural network (DNN) is a central challenge in machine learning. Several proposed local explanation methods address this issue by identifying what dimensions of a single input are most responsible for a DNN's output. The goal of this work is to assess the sensitivity of local explanations to DNN parameter values. Somewhat surprisingly, we find that DNNs with randomly-initialized weights produce explanations that are both visually and quantitatively similar to those produced by DNNs with learned weights. Our conjecture is that this phenomenon occurs because these explanations are dominated by the lower level features of a DNN, and that a DNN's architecture provides a strong prior which significantly affects the representations learned at these lower layers. NOTE: This work is now subsumed by our recent manuscript, Sanity Checks for Saliency Maps (to appear NIPS 2018), where we expand on findings and address concerns raised in Sundararajan et. al. (2018).

READ FULL TEXT

page 2

page 7

page 8

page 9

page 10

research
06/11/2018

A Note about: Local Explanation Methods for Deep Neural Networks lack Sensitivity to Parameter Values

Local explanation methods, also known as attribution methods, attribute ...
research
11/09/2022

On the Robustness of Explanations of Deep Neural Network Models: A Survey

Explainability has been widely stated as a cornerstone of the responsibl...
research
10/25/2022

Towards Formal XAI: Formally Approximate Minimal Explanations of Neural Networks

With the rapid growth of machine learning, deep neural networks (DNNs) a...
research
01/27/2019

How Sensitive are Sensitivity-Based Explanations?

We propose a simple objective evaluation measure for explanations of a c...
research
04/26/2023

Technical Note: Defining and Quantifying AND-OR Interactions for Faithful and Concise Explanation of DNNs

In this technical note, we aim to explain a deep neural network (DNN) by...
research
03/20/2018

Explanation Methods in Deep Learning: Users, Values, Concerns and Challenges

Issues regarding explainable AI involve four components: users, laws & r...
research
07/31/2023

Formally Explaining Neural Networks within Reactive Systems

Deep neural networks (DNNs) are increasingly being used as controllers i...

Please sign up or login with your details

Forgot password? Click here to reset