Shortcut Detection with Variational Autoencoders

02/08/2023
by   Nicolas M. Müller, et al.
0

For real-world applications of machine learning (ML), it is essential that models make predictions based on well-generalizing features rather than spurious correlations in the data. The identification of such spurious correlations, also known as shortcuts, is a challenging problem and has so far been scarcely addressed. In this work, we present a novel approach to detect shortcuts in image and audio datasets by leveraging variational autoencoders (VAEs). The disentanglement of features in the latent space of VAEs allows us to discover correlations in datasets and semi-automatically evaluate them for ML shortcuts. We demonstrate the applicability of our method on several real-world datasets and identify shortcuts that have not been discovered before. Based on these findings, we also investigate the construction of shortcut adversarial examples.

READ FULL TEXT

page 5

page 7

page 8

page 12

page 13

research
07/05/2023

On the Adversarial Robustness of Generative Autoencoders in the Latent Space

The generative autoencoders, such as the variational autoencoders or the...
research
07/16/2021

Machine-learning Kondo physics using variational autoencoders

We employ variational autoencoders to extract physical insight from a da...
research
02/18/2022

Rethinking Machine Learning Robustness via its Link with the Out-of-Distribution Problem

Despite multiple efforts made towards robust machine learning (ML) model...
research
08/03/2021

On the Exploitability of Audio Machine Learning Pipelines to Surreptitious Adversarial Examples

Machine learning (ML) models are known to be vulnerable to adversarial e...
research
10/12/2020

Anomaly Detection With Conditional Variational Autoencoders

Exploiting the rapid advances in probabilistic inference, in particular ...
research
09/15/2023

Quantifying Credit Portfolio sensitivity to asset correlations with interpretable generative neural networks

In this research, we propose a novel approach for the quantification of ...
research
07/22/2022

TRUST-LAPSE: An Explainable Actionable Mistrust Scoring Framework for Model Monitoring

Continuous monitoring of trained ML models to determine when their predi...

Please sign up or login with your details

Forgot password? Click here to reset