Learning Important Features Through Propagating Activation Differences

04/10/2017
by   Avanti Shrikumar, et al.
0

The purported "black box"' nature of neural networks is a barrier to adoption in applications where interpretability is essential. Here we present DeepLIFT (Deep Learning Important FeaTures), a method for decomposing the output prediction of a neural network on a specific input by backpropagating the contributions of all neurons in the network to every feature of the input. DeepLIFT compares the activation of each neuron to its 'reference activation' and assigns contribution scores according to the difference. By optionally giving separate consideration to positive and negative contributions, DeepLIFT can also reveal dependencies which are missed by other approaches. Scores can be computed efficiently in a single backward pass. We apply DeepLIFT to models trained on MNIST and simulated genomic data, and show significant advantages over gradient-based methods. A detailed video tutorial on the method is at http://goo.gl/qKb7pL and code is at http://goo.gl/RM8jvH.

READ FULL TEXT

page 6

page 7

research
05/05/2016

Not Just a Black Box: Learning Important Features Through Propagating Activation Differences

Note: This paper describes an older version of DeepLIFT. See https://arx...
research
11/25/2019

Explaining Neural Networks via Perturbing Important Learned Features

Attributing the output of a neural network to the contribution of given ...
research
07/11/2023

Using Linear Regression for Iteratively Training Neural Networks

We present a simple linear regression based approach for learning the we...
research
04/01/2019

Relative Attributing Propagation: Interpreting the Comparative Contributions of Individual Units in Deep Neural Networks

As Deep Neural Networks (DNNs) have demonstrated superhuman performance ...
research
08/20/2020

Prototype-based interpretation of the functionality of neurons in winner-take-all neural networks

Prototype-based learning (PbL) using a winner-take-all (WTA) network bas...
research
05/21/2021

Condition Integration Memory Network: An Interpretation of the Meaning of the Neuronal Design

This document introduces a hypothesized framework on the functional natu...

Please sign up or login with your details

Forgot password? Click here to reset