Measurably Stronger Explanation Reliability via Model Canonization

02/14/2022
by   Franz Motzkus, et al.
0

While rule-based attribution methods have proven useful for providing local explanations for Deep Neural Networks, explaining modern and more varied network architectures yields new challenges in generating trustworthy explanations, since the established rule sets might not be sufficient or applicable to novel network structures. As an elegant solution to the above issue, network canonization has recently been introduced. This procedure leverages the implementation-dependency of rule-based attributions and restructures a model into a functionally identical equivalent of alternative design to which established attribution rules can be applied. However, the idea of canonization and its usefulness have so far only been explored qualitatively. In this work, we quantitatively verify the beneficial effects of network canonization to rule-based attributions on VGG-16 and ResNet18 models with BatchNorm layers and thus extend the current best practices for obtaining reliable neural network explanations.

READ FULL TEXT

page 3

page 4

research
01/22/2023

The Shape of Explanations: A Topological Account of Rule-Based Explanations in Machine Learning

Rule-based explanations provide simple reasons explaining the behavior o...
research
11/30/2022

Optimizing Explanations by Network Canonization and Hyperparameter Search

Explainable AI (XAI) is slowly becoming a key component for many AI appl...
research
10/24/2022

Generating Hierarchical Explanations on Text Classification Without Connecting Rules

The opaqueness of deep NLP models has motivated the development of metho...
research
08/04/2020

Explanation of Reinforcement Learning Model in Dynamic Multi-Agent System

Recently, there has been increasing interest in transparency and interpr...
research
04/11/2023

CGXplain: Rule-Based Deep Neural Network Explanations Using Dual Linear Programs

Rule-based surrogate models are an effective and interpretable way to ap...
research
06/22/2017

Explaining Recurrent Neural Network Predictions in Sentiment Analysis

Recently, a technique called Layer-wise Relevance Propagation (LRP) was ...
research
09/06/2023

A Refutation of Shapley Values for Explainability

Recent work demonstrated the existence of Boolean functions for which Sh...

Please sign up or login with your details

Forgot password? Click here to reset