Don't Throw it Away! The Utility of Unlabeled Data in Fair Decision Making

05/10/2022
by   Miriam Rateike, et al.
31

Decision making algorithms, in practice, are often trained on data that exhibits a variety of biases. Decision-makers often aim to take decisions based on some ground-truth target that is assumed or expected to be unbiased, i.e., equally distributed across socially salient groups. In many practical settings, the ground-truth cannot be directly observed, and instead, we have to rely on a biased proxy measure of the ground-truth, i.e., biased labels, in the data. In addition, data is often selectively labeled, i.e., even the biased labels are only observed for a small fraction of the data that received a positive decision. To overcome label and selection biases, recent work proposes to learn stochastic, exploring decision policies via i) online training of new policies at each time-step and ii) enforcing fairness as a constraint on performance. However, the existing approach uses only labeled data, disregarding a large amount of unlabeled data, and thereby suffers from high instability and variance in the learned decision policies at different times. In this paper, we propose a novel method based on a variational autoencoder for practical fair decision-making. Our method learns an unbiased data representation leveraging both labeled and unlabeled data and uses the representations to learn a policy in an online process. Using synthetic data, we empirically validate that our method converges to the optimal (fair) policy according to the ground-truth with low variance. In real-world experiments, we further show that our training approach not only offers a more stable learning process but also yields policies with higher fairness as well as utility than previous approaches.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/08/2021

A Ranking Approach to Fair Classification

Algorithmic decision systems are increasingly used in areas such as hiri...
research
02/08/2019

Improving Consequential Decision Making under Imperfect Predictions

Consequential decisions are increasingly informed by sophisticated data-...
research
05/15/2023

MADDM: Multi-Advisor Dynamic Binary Decision-Making by Maximizing the Utility

Being able to infer ground truth from the responses of multiple imperfec...
research
05/02/2022

A Novel Approach to Fairness in Automated Decision-Making using Affective Normalization

Any decision, such as one about who to hire, involves two components. Fi...
research
02/09/2022

Designing Closed Human-in-the-loop Deferral Pipelines

In hybrid human-machine deferral frameworks, a classifier can defer unce...
research
09/30/2020

Uncertainty Estimation For Community Standards Violation In Online Social Networks

Online Social Networks (OSNs) provide a platform for users to share thei...
research
03/30/2023

Mitigating Source Bias for Fairer Weak Supervision

Weak supervision overcomes the label bottleneck, enabling efficient deve...

Please sign up or login with your details

Forgot password? Click here to reset