Explaining Deep Convolutional Neural Networks via Latent Visual-Semantic Filter Attention

04/10/2022
by   Yu Yang, et al.
0

Interpretability is an important property for visual models as it helps researchers and users understand the internal mechanism of a complex model. However, generating semantic explanations about the learned representation is challenging without direct supervision to produce such explanations. We propose a general framework, Latent Visual Semantic Explainer (LaViSE), to teach any existing convolutional neural network to generate text descriptions about its own latent representations at the filter level. Our method constructs a mapping between the visual and semantic spaces using generic image datasets, using images and category names. It then transfers the mapping to the target domain which does not have semantic labels. The proposed framework employs a modular structure and enables to analyze any trained network whether or not its original training data is available. We show that our method can generate novel descriptions for learned filters beyond the set of categories defined in the training dataset and perform an extensive evaluation on multiple datasets. We also demonstrate a novel application of our method for unsupervised dataset bias analysis which allows us to automatically discover hidden biases in datasets or compare different subsets without using additional labels. The dataset and code are made public to facilitate further research.

READ FULL TEXT

page 3

page 6

page 7

page 8

page 11

research
11/15/2017

Interpreting Deep Visual Representations via Network Dissection

The success of recent deep convolutional neural networks (CNNs) depends ...
research
01/30/2023

NeSyFOLD: A System for Generating Logic-based Explanations from Convolutional Neural Networks

We present a novel neurosymbolic system called NeSyFOLD that classifies ...
research
04/19/2017

Network Dissection: Quantifying Interpretability of Deep Visual Representations

We propose a general framework called Network Dissection for quantifying...
research
04/27/2020

A Disentangling Invertible Interpretation Network for Explaining Latent Representations

Neural networks have greatly boosted performance in computer vision by l...
research
06/29/2015

Tell and Predict: Kernel Classifier Prediction for Unseen Visual Classes from Unstructured Text Descriptions

In this paper we propose a framework for predicting kernelized classifie...
research
11/18/2019

Towards Visually Explaining Variational Autoencoders

Recent advances in Convolutional Neural Network (CNN) model interpretabi...
research
10/21/2019

Boosting Mapping Functionality of Neural Networks via Latent Feature Generation based on Reversible Learning

This paper addresses a boosting method for mapping functionality of neur...

Please sign up or login with your details

Forgot password? Click here to reset