Improving Interpretability of Deep Neural Networks with Semantic Information

03/12/2017
by   Yinpeng Dong, et al.
0

Interpretability of deep neural networks (DNNs) is essential since it enables users to understand the overall strengths and weaknesses of the models, conveys an understanding of how the models will behave in the future, and how to diagnose and correct potential problems. However, it is challenging to reason about what a DNN actually does due to its opaque or black-box nature. To address this issue, we propose a novel technique to improve the interpretability of DNNs by leveraging the rich semantic information embedded in human descriptions. By concentrating on the video captioning task, we first extract a set of semantically meaningful topics from the human descriptions that cover a wide range of visual concepts, and integrate them into the model with an interpretive loss. We then propose a prediction difference maximization algorithm to interpret the learned features of each neuron. Experimental results demonstrate its effectiveness in video captioning using the interpretable features, which can also be transferred to video action recognition. By clearly understanding the learned features, users can easily revise false predictions via a human-in-the-loop procedure.

READ FULL TEXT

page 7

page 8

research
09/12/2019

New Perspective of Interpretability of Deep Neural Networks

Deep neural networks (DNNs) are known as black-box models. In other word...
research
10/27/2020

Quantifying Learnability and Describability of Visual Concepts Emerging in Representation Learning

The increasing impact of black box models, and particularly of unsupervi...
research
10/21/2019

Semantics for Global and Local Interpretation of Deep Neural Networks

Deep neural networks (DNNs) with high expressiveness have achieved state...
research
01/25/2019

Towards Interpretable Deep Neural Networks by Leveraging Adversarial Examples

Sometimes it is not enough for a DNN to produce an outcome. For example,...
research
06/26/2017

Cognitive Psychology for Deep Neural Networks: A Shape Bias Case Study

Deep neural networks (DNNs) have achieved unprecedented performance on a...
research
10/22/2020

Towards falsifiable interpretability research

Methods for understanding the decisions of and mechanisms underlying dee...
research
09/07/2023

A Function Interpretation Benchmark for Evaluating Interpretability Methods

Labeling neural network submodules with human-legible descriptions is us...

Please sign up or login with your details

Forgot password? Click here to reset