Learning Interpretable Concept-Based Models with Human Feedback

12/04/2020
by   Isaac Lage, et al.
0

Machine learning models that first learn a representation of a domain in terms of human-understandable concepts, then use it to make predictions, have been proposed to facilitate interpretation and interaction with models trained on high-dimensional data. However these methods have important limitations: the way they define concepts are not inherently interpretable, and they assume that concept labels either exist for individual instances or can easily be acquired from users. These limitations are particularly acute for high-dimensional tabular features. We propose an approach for learning a set of transparent concept definitions in high-dimensional tabular data that relies on users labeling concept features instead of individual instances. Our method produces concepts that both align with users' intuitive sense of what a concept means, and facilitate prediction of the downstream label by a transparent machine learning model. This ensures that the full model is transparent and intuitive, and as predictive as possible given this constraint. We demonstrate with simulated user feedback on real prediction problems, including one in a clinical domain, that this kind of direct feedback is much more efficient at learning solutions that align with ground truth concept definitions than alternative transparent approaches that rely on labeling instances or other existing interaction mechanisms, while maintaining similar predictive performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/07/2022

TCNL: Transparent and Controllable Network Learning Via Embedding Human-Guided Concepts

Explaining deep learning models is of vital importance for understanding...
research
03/29/2018

Proof-of-Concept Examples of Performance-Transparent Programming Models

Machine-specific optimizations command the machine to behave in a specif...
research
12/14/2022

Interactive Concept Bottleneck Models

Concept bottleneck models (CBMs) (Koh et al. 2020) are interpretable neu...
research
04/27/2023

Interpretable Neural-Symbolic Concept Reasoning

Deep learning methods are highly accurate, yet their opaque decision pro...
research
10/18/2016

An Interactive Machine Learning Framework

Machine learning (ML) is believed to be an effective and efficient tool ...
research
05/27/2021

Intellige: A User-Facing Model Explainer for Narrative Explanations

Predictive machine learning models often lack interpretability, resultin...
research
08/08/2021

Human-in-the-loop Extraction of Interpretable Concepts in Deep Learning Models

The interpretation of deep neural networks (DNNs) has become a key topic...

Please sign up or login with your details

Forgot password? Click here to reset