NeuroMask: Explaining Predictions of Deep Neural Networks through Mask Learning

08/05/2019
by   Moustafa Alzantot, et al.
6

Deep Neural Networks (DNNs) deliver state-of-the-art performance in many image recognition and understanding applications. However, despite their outstanding performance, these models are black-boxes and it is hard to understand how they make their decisions. Over the past few years, researchers have studied the problem of providing explanations of why DNNs predicted their results. However, existing techniques are either obtrusive, requiring changes in model training, or suffer from low output quality. In this paper, we present a novel method, NeuroMask, for generating an interpretable explanation of classification model results. When applied to image classification models, NeuroMask identifies the image parts that are most important to classifier results by applying a mask that hides/reveals different parts of the image, before feeding it back into the model. The mask values are tuned by minimizing a properly designed cost function that preserves the classification result and encourages producing an interpretable mask. Experiments using state-of-the-art Convolutional Neural Networks for image recognition on different datasets (CIFAR-10 and ImageNet) show that NeuroMask successfully localizes the parts of the input image which are most relevant to the DNN decision. By showing a visual quality comparison between NeuroMask explanations and those of other methods, we find NeuroMask to be both accurate and interpretable.

READ FULL TEXT

page 1

page 3

page 5

page 6

research
11/27/2022

Foiling Explanations in Deep Neural Networks

Deep neural networks (DNNs) have greatly impacted numerous fields over t...
research
10/17/2019

Effect of Superpixel Aggregation on Explanations in LIME – A Case Study with Biological Data

End-to-end learning with deep neural networks, such as convolutional neu...
research
11/22/2022

Explaining Image Classifiers with Multiscale Directional Image Representation

Image classifiers are known to be difficult to interpret and therefore r...
research
12/08/2015

Explaining NonLinear Classification Decisions with Deep Taylor Decomposition

Nonlinear methods such as Deep Neural Networks (DNNs) are the gold stand...
research
08/20/2019

Saccader: Improving Accuracy of Hard Attention Models for Vision

Although deep convolutional neural networks achieve state-of-the-art per...
research
10/06/2021

Disentangling deep neural networks with rectified linear units using duality

Despite their success deep neural networks (DNNs) are still largely cons...
research
07/01/2022

Offset equivariant networks and their applications

In this paper we present a framework for the design and implementation o...

Please sign up or login with your details

Forgot password? Click here to reset