On the impossibility of non-trivial accuracy under fairness constraints

07/14/2021
by   Carlos Pinzón, et al.
0

One of the main concerns about fairness in machine learning (ML) is that, in order to achieve it, one may have to renounce to some accuracy. Having this trade-off in mind, Hardt et al. have proposed the notion of equal opportunities (EO), designed so as to be compatible with accuracy. In fact, it can be shown that if the source of input data is deterministic, the two notions go well along with each other. In the probabilistic case, however, things change. As we show, there are probabilistic data sources for which EO can only be achieved at the total detriment of accuracy, i.e. among the models that achieve EO, those whose prediction does not depend on the input have the highest accuracy.

READ FULL TEXT VIEW PDF
POST COMMENT

Comments

There are no comments yet.

Authors

page 1

page 2

page 3

page 4

08/13/2021

The Sharpe predictor for fairness in machine learning

In machine learning (ML) applications, unfair predictions may discrimina...
02/01/2021

Emergent Unfairness: Normative Assumptions and Contradictions in Algorithmic Fairness-Accuracy Trade-Off Research

Across machine learning (ML) sub-disciplines, researchers make explicit ...
08/24/2018

An Empirical Study of Rich Subgroup Fairness for Machine Learning

Kearns et al. [2018] recently proposed a notion of rich subgroup fairnes...
08/05/2021

Reducing Unintended Bias of ML Models on Tabular and Textual Data

Unintended biases in machine learning (ML) models are among the major co...
10/07/2020

Fairness in Influence Maximization through Randomization

The influence maximization paradigm has been used by researchers in vari...
06/09/2020

Fair Bayesian Optimization

Given the increasing importance of machine learning (ML) in our lives, a...
This week in AI

Get the week's most popular data science and artificial intelligence research sent straight to your inbox every Saturday.

Appendix A Appendix

[body]lemmas.tex