Evaluating the Interpretability of Generative Models by Interactive Reconstruction

02/02/2021
by   Andrew Slavin Ross, et al.
13

For machine learning models to be most useful in numerous sociotechnical systems, many have argued that they must be human-interpretable. However, despite increasing interest in interpretability, there remains no firm consensus on how to measure it. This is especially true in representation learning, where interpretability research has focused on "disentanglement" measures only applicable to synthetic datasets and not grounded in human factors. We introduce a task to quantify the human-interpretability of generative model representations, where users interactively modify representations to reconstruct target instances. On synthetic datasets, we find performance on this task much more reliably differentiates entangled and disentangled models than baseline approaches. On a real dataset, we find it differentiates between representation learning methods widely believed but never shown to produce more or less interpretable models. In both cases, we ran small-scale think-aloud studies and large-scale experiments on Amazon Mechanical Turk to confirm that our qualitative and quantitative results agreed.

READ FULL TEXT

page 5

page 7

page 18

page 20

research
04/06/2023

DSVAE: Interpretable Disentangled Representation for Synthetic Speech Detection

Tools to generate high quality synthetic speech signal that is perceptua...
research
08/14/2021

Unsupervised Disentanglement without Autoencoding: Pitfalls and Future Directions

Disentangled visual representations have largely been studied with gener...
research
07/08/2019

The Price of Interpretability

When quantitative models are used to support decision-making on complex ...
research
02/21/2018

Manipulating and Measuring Model Interpretability

Despite a growing body of research focused on creating interpretable mac...
research
06/05/2020

Evaluating the Disentanglement of Deep Generative Models through Manifold Topology

Learning disentangled representations is regarded as a fundamental task ...
research
07/11/2023

Scale Alone Does not Improve Mechanistic Interpretability in Vision Models

In light of the recent widespread adoption of AI systems, understanding ...
research
11/22/2020

A Bayesian Account of Measures of Interpretability in Human-AI Interaction

Existing approaches for the design of interpretable agent behavior consi...

Please sign up or login with your details

Forgot password? Click here to reset