Explaining Explanations: Axiomatic Feature Interactions for Deep Networks

02/10/2020
by   Joseph D. Janizek, et al.
30

Recent work has shown great promise in explaining neural network behavior. In particular, feature attribution methods explain which features were most important to a model's prediction on a given input. However, for many tasks, simply knowing which features were important to a model's prediction may not provide enough insight to understand model behavior. The interactions between features within the model may better help us understand not only the model, but also why certain features are more important than others. In this work, we present Integrated Hessians (Code available at https://github.com/suinleelab/path_explain): an extension of Integrated Gradients that explains pairwise feature interactions in neural networks. Integrated Hessians overcomes several theoretical limitations of previous methods to explain interactions, and unlike such previous methods is not limited to a specific architecture or class of neural network. We apply Integrated Hessians on a variety of neural networks trained on language data, biological data, astronomy data, and medical data and gain new insight into model behavior in each domain.

READ FULL TEXT

page 4

page 6

page 21

page 22

page 23

page 24

page 28

research
05/22/2022

Visual Explanations from Deep Networks via Riemann-Stieltjes Integrated Gradient-based Localization

Neural networks are becoming increasingly better at tasks that involve c...
research
08/29/2018

Rule induction for global explanation of trained models

Understanding the behavior of a trained network and finding explanations...
research
11/08/2016

Gradients of Counterfactuals

Gradients have been used to quantify feature importance in machine learn...
research
03/08/2016

A New Method to Visualize Deep Neural Networks

We present a method for visualising the response of a deep neural networ...
research
09/04/2019

Generalized Integrated Gradients: A practical method for explaining diverse ensembles

We introduce Generalized Integrated Gradients (GIG), a formal extension ...
research
08/08/2023

ScatterUQ: Interactive Uncertainty Visualizations for Multiclass Deep Learning Problems

Recently, uncertainty-aware deep learning methods for multiclass labelin...
research
02/15/2021

Unified Shapley Framework to Explain Prediction Drift

Predictions are the currency of a machine learning model, and to underst...

Please sign up or login with your details

Forgot password? Click here to reset