TRUST-LAPSE: An Explainable Actionable Mistrust Scoring Framework for Model Monitoring

07/22/2022
by   Nandita Bhaskhar, et al.
16

Continuous monitoring of trained ML models to determine when their predictions should and should not be trusted is essential for their safe deployment. Such a framework ought to be high-performing, explainable, post-hoc and actionable. We propose TRUST-LAPSE, a "mistrust" scoring framework for continuous model monitoring. We assess the trustworthiness of each input sample's model prediction using a sequence of latent-space embeddings. Specifically, (a) our latent-space mistrust score estimates mistrust using distance metrics (Mahalanobis distance) and similarity metrics (cosine similarity) in the latent-space and (b) our sequential mistrust score determines deviations in correlations over the sequence of past input representations in a non-parametric, sliding-window based algorithm for actionable continuous monitoring. We evaluate TRUST-LAPSE via two downstream tasks: (1) distributionally shifted input detection and (2) data drift detection, across diverse domains – audio vision using public datasets and further benchmark our approach on challenging, real-world electroencephalograms (EEG) datasets for seizure detection. Our latent-space mistrust scores achieve state-of-the-art results with AUROCs of 84.1 (vision), 73.9 (audio), 77.1 (clinical EEGs), outperforming baselines by over 10 points. We expose critical failures in popular baselines that remain insensitive to input semantic content, rendering them unfit for real-world model monitoring. We show that our sequential mistrust scores achieve high drift detection rates: over 90 streams show < 20 quantitative evaluations, we show that our mistrust scores are more robust and provide explainability for easy adoption into practice.

READ FULL TEXT

page 1

page 19

research
10/19/2020

Evidential Sparsification of Multimodal Latent Spaces in Conditional Variational Autoencoders

Discrete latent spaces in variational autoencoders have been shown to ef...
research
10/08/2018

Towards the Latent Transcriptome

In this work we propose a method to compute continuous embeddings for km...
research
09/22/2022

Assessing Robustness of EEG Representations under Data-shifts via Latent Space and Uncertainty Analysis

The recent availability of large datasets in bio-medicine has inspired t...
research
06/16/2023

Prototype Learning for Explainable Regression

The lack of explainability limits the adoption of deep learning models i...
research
02/08/2023

Shortcut Detection with Variational Autoencoders

For real-world applications of machine learning (ML), it is essential th...
research
12/28/2022

Exploration of latent space of LOD2 GML dataset to identify similar buildings

Explainable numerical representations of otherwise complex datasets are ...

Please sign up or login with your details

Forgot password? Click here to reset