Explaining NonLinear Classification Decisions with Deep Taylor Decomposition

12/08/2015
by   Grégoire Montavon, et al.
0

Nonlinear methods such as Deep Neural Networks (DNNs) are the gold standard for various challenging machine learning problems, e.g., image classification, natural language processing or human action recognition. Although these methods perform impressively well, they have a significant disadvantage, the lack of transparency, limiting the interpretability of the solution and thus the scope of application in practice. Especially DNNs act as black boxes due to their multilayer nonlinear structure. In this paper we introduce a novel methodology for interpreting generic multilayer neural networks by decomposing the network classification decision into contributions of its input elements. Although our focus is on image classification, the method is applicable to a broad set of input data, learning tasks and network architectures. Our method is based on deep Taylor decomposition and efficiently utilizes the structure of the network by backpropagating the explanations from the output to the input layer. We evaluate the proposed method empirically on the MNIST and ILSVRC data sets.

READ FULL TEXT

page 8

page 11

page 12

page 19

page 20

research
12/03/2018

Sensitivity based Neural Networks Explanations

Although neural networks can achieve very high predictive performance on...
research
04/21/2019

Explaining a prediction in some nonlinear models

In this article we will analyse how to compute the contribution of each ...
research
08/05/2019

NeuroMask: Explaining Predictions of Deep Neural Networks through Mask Learning

Deep Neural Networks (DNNs) deliver state-of-the-art performance in many...
research
07/09/2018

Interpreting and Explaining Deep Neural Networks for Classification of Audio Signals

Interpretability of deep neural networks is a recently emerging area of ...
research
02/13/2023

A Domain Decomposition-Based CNN-DNN Architecture for Model Parallel Training Applied to Image Recognition Problems

Deep neural networks (DNNs) and, in particular, convolutional neural net...
research
01/28/2019

Interpreting Deep Neural Networks Through Variable Importance

While the success of deep neural networks (DNNs) is well-established acr...
research
07/15/2019

A study on the Interpretability of Neural Retrieval Models using DeepSHAP

A recent trend in IR has been the usage of neural networks to learn retr...

Please sign up or login with your details

Forgot password? Click here to reset