Multiaccurate Proxies for Downstream Fairness

07/09/2021
by   Emily Diana, et al.
1

We study the problem of training a model that must obey demographic fairness conditions when the sensitive features are not available at training time – in other words, how can we train a model to be fair by race when we don't have data about race? We adopt a fairness pipeline perspective, in which an "upstream" learner that does have access to the sensitive features will learn a proxy model for these features from the other attributes. The goal of the proxy is to allow a general "downstream" learner – with minimal assumptions on their prediction task – to be able to use the proxy to train a model that is fair with respect to the true sensitive features. We show that obeying multiaccuracy constraints with respect to the downstream model class suffices for this purpose, and provide sample- and oracle efficient-algorithms and generalization bounds for learning such proxies. In general, multiaccuracy can be much easier to satisfy than classification accuracy, and can be satisfied even when the sensitive features are hard to predict.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/26/2023

Balanced Filtering via Non-Disclosive Proxies

We study the problem of non-disclosively collecting a sample of data tha...
research
07/25/2022

Estimating and Controlling for Fairness via Sensitive Attribute Predictors

Although machine learning classifiers have been increasingly used in hig...
research
08/12/2021

Fair Decision-Making for Food Inspections

We revisit the application of predictive models by the Chicago Departmen...
research
05/20/2023

Model Debiasing via Gradient-based Explanation on Representation

Machine learning systems produce biased results towards certain demograp...
research
11/02/2022

Fair Visual Recognition via Intervention with Proxy Features

Deep learning models often learn to make predictions that rely on sensit...
research
10/31/2018

Crowdsourcing with Fairness, Diversity and Budget Constraints

Recent studies have shown that the labels collected from crowdworkers ca...
research
01/30/2019

Noise-tolerant fair classification

Fair machine learning concerns the analysis and design of learning algor...

Please sign up or login with your details

Forgot password? Click here to reset