Aleatoric and Epistemic Discrimination in Classification

01/27/2023
by   Hao Wang, et al.
0

Machine learning (ML) models can underperform on certain population groups due to choices made during model development and bias inherent in the data. We categorize sources of discrimination in the ML pipeline into two classes: aleatoric discrimination, which is inherent in the data distribution, and epistemic discrimination, which is due to decisions during model development. We quantify aleatoric discrimination by determining the performance limits of a model under fairness constraints, assuming perfect knowledge of the data distribution. We demonstrate how to characterize aleatoric discrimination by applying Blackwell's results on comparing statistical experiments. We then quantify epistemic discrimination as the gap between a model's accuracy given fairness constraints and the limit posed by aleatoric discrimination. We apply this approach to benchmark existing interventions and investigate fairness risks in data with missing values. Our results indicate that state-of-the-art fairness interventions are effective at removing epistemic discrimination. However, when data has missing values, there is still significant room for improvement in handling aleatoric discrimination.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/30/2023

Adapting Fairness Interventions to Missing Values

Missing values in real-world data pose a significant and unique challeng...
research
06/08/2023

Shedding light on underrepresentation and Sampling Bias in machine learning

Accurately measuring discrimination is crucial to faithfully assessing f...
research
07/14/2020

A Normative approach to Attest Digital Discrimination

Digital discrimination is a form of discrimination whereby users are aut...
research
11/27/2020

Black Loans Matter: Distributionally Robust Fairness for Fighting Subgroup Discrimination

Algorithmic fairness in lending today relies on group fairness metrics f...
research
09/11/2017

Fairness Testing: Testing Software for Discrimination

This paper defines software fairness and discrimination and develops a t...
research
11/29/2021

A Causal Approach for Unfair Edge Prioritization and Discrimination Removal

In budget-constrained settings aimed at mitigating unfairness, like law ...
research
08/24/2022

TESTSGD: Interpretable Testing of Neural Networks Against Subtle Group Discrimination

Discrimination has been shown in many machine learning applications, whi...

Please sign up or login with your details

Forgot password? Click here to reset