Human-in-the-loop Extraction of Interpretable Concepts in Deep Learning Models

08/08/2021
by   Zhenge Zhao, et al.
0

The interpretation of deep neural networks (DNNs) has become a key topic as more and more people apply them to solve various problems and making critical decisions. Concept-based explanations have recently become a popular approach for post-hoc interpretation of DNNs. However, identifying human-understandable visual concepts that affect model decisions is a challenging task that is not easily addressed with automatic approaches. We present a novel human-in-the-loop approach to generate user-defined concepts for model interpretation and diagnostics. Central to our proposal is the use of active learning, where human knowledge and feedback are combined to train a concept extractor with very little human labeling effort. We integrate this process into an interactive system, ConceptExtract. Through two case studies, we show how our approach helps analyze model behavior and extract human-friendly concepts for different machine learning tasks and datasets and how to use these concepts to understand the predictions, compare model performance and make suggestions for model refinement. Quantitative experiments show that our active learning approach can accurately extract meaningful visual concepts. More importantly, by identifying visual concepts that negatively affect model performance, we develop the corresponding data augmentation strategy that consistently improves model performance.

READ FULL TEXT

page 1

page 4

page 6

page 8

research
10/07/2019

A Survey on Active Learning and Human-in-the-Loop Deep Learning for Medical Image Analysis

Fully automatic deep learning has become the state-of-the-art technique ...
research
08/01/2019

Semantic Concept Spaces: Guided Topic Model Refinement using Word-Embedding Projections

We present a framework that allows users to incorporate the semantics of...
research
05/31/2016

Towards ontology driven learning of visual concept detectors

The maturity of deep learning techniques has led in recent years to a br...
research
06/13/2022

Efficient Human-in-the-loop System for Guiding DNNs Attention

Attention guidance is an approach to addressing dataset bias in deep lea...
research
07/10/2023

Hierarchical Semantic Tree Concept Whitening for Interpretable Image Classification

With the popularity of deep neural networks (DNNs), model interpretabili...
research
12/04/2020

Learning Interpretable Concept-Based Models with Human Feedback

Machine learning models that first learn a representation of a domain in...
research
03/06/2023

NxPlain: Web-based Tool for Discovery of Latent Concepts

The proliferation of deep neural networks in various domains has seen an...

Please sign up or login with your details

Forgot password? Click here to reset