Group Fairness by Probabilistic Modeling with Latent Fair Decisions

09/18/2020
by   YooJung Choi, et al.
29

Machine learning systems are increasingly being used to make impactful decisions such as loan applications and criminal justice risk assessments, and as such, ensuring fairness of these systems is critical. This is often challenging as the labels in the data are biased. This paper studies learning fair probability distributions from biased data by explicitly modeling a latent variable that represents a hidden, unbiased label. In particular, we aim to achieve demographic parity by enforcing certain independencies in the learned model. We also show that group fairness guarantees are meaningful only if the distribution used to provide those guarantees indeed captures the real-world data. In order to closely model the data distribution, we employ probabilistic circuits, an expressive and tractable probabilistic model, and propose an algorithm to learn them from incomplete data. We evaluate our approach on a synthetic dataset in which observed labels indeed come from fair labels but with added bias, and demonstrate that the fair labels are successfully retrieved. Moreover, we show on real-world datasets that our approach not only is a better model than existing methods of how the data was generated but also achieves competitive accuracy.

READ FULL TEXT
POST COMMENT

Comments

There are no comments yet.

Authors

page 6

03/20/2017

Counterfactual Fairness

Machine learning can impact people with legal or ethical consequences wh...
07/07/2021

Bias-Tolerant Fair Classification

The label bias and selection bias are acknowledged as two reasons in dat...
01/15/2019

Identifying and Correcting Label Bias in Machine Learning

Datasets often contain biases which unfairly disadvantage certain groups...
11/29/2021

Learning Fair Classifiers with Partially Annotated Group Labels

Recently, fairness-aware learning have become increasingly crucial, but ...
12/01/2020

Data Preprocessing to Mitigate Bias with Boosted Fair Mollifiers

In a recent paper, Celis et al. (2020) introduced a new approach to fair...
05/10/2022

Don't Throw it Away! The Utility of Unlabeled Data in Fair Decision Making

Decision making algorithms, in practice, are often trained on data that ...
01/24/2019

Pretending Fair Decisions via Stealthily Biased Sampling

Fairness by decision-makers is believed to be auditable by third parties...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.