On Codomain Separability and Label Inference from (Noisy) Loss Functions

07/07/2021
by   Abhinav Aggarwal, et al.
1

Machine learning classifiers rely on loss functions for performance evaluation, often on a private (hidden) dataset. Label inference was recently introduced as the problem of reconstructing the ground truth labels of this private dataset from just the (possibly perturbed) loss function values evaluated at chosen prediction vectors, without any other access to the hidden dataset. Existing results have demonstrated this inference is possible on specific loss functions like the cross-entropy loss. In this paper, we introduce the notion of codomain separability to formally study the necessary and sufficient conditions under which label inference is possible from any (noisy) loss function values. Using this notion, we show that for many commonly used loss functions, including multiclass cross-entropy with common activation functions and some Bregman divergence-based losses, it is possible to design label inference attacks for arbitrary noise levels. We demonstrate that these attacks can also be carried out through actual neural network models, and argue, both formally and empirically, the role of finite precision arithmetic in this setting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/08/2023

Reevaluating Loss Functions: Enhancing Robustness to Label Noise in Deep Learning Models

Large annotated datasets inevitably contain incorrect labels, which pose...
research
05/18/2021

Label Inference Attacks from Log-loss Scores

Log-loss (also known as cross-entropy loss) metric is ubiquitously used ...
research
03/08/2023

Unimodal Distributions for Ordinal Regression

In many real-world prediction tasks, class labels contain information ab...
research
05/10/2021

Generalized Jensen-Shannon Divergence Loss for Learning with Noisy Labels

We propose two novel loss functions based on Jensen-Shannon divergence f...
research
06/07/2023

Loss Functions for Behavioral Game Theory

Behavioral game theorists all use experimental data to evaluate predicti...
research
03/06/2023

The α-divergence Improves the Entropy Production Estimation via Machine Learning

Recent years have seen a surge of interest in the algorithmic estimation...
research
10/10/2018

Complementary-Label Learning for Arbitrary Losses and Models

In contrast to the standard classification paradigm where the true (or p...

Please sign up or login with your details

Forgot password? Click here to reset