Does Data Repair Lead to Fair Models? Curating Contextually Fair Data To Reduce Model Bias

10/20/2021
by   Sharat Agarwal, et al.
0

Contextual information is a valuable cue for Deep Neural Networks (DNNs) to learn better representations and improve accuracy. However, co-occurrence bias in the training dataset may hamper a DNN model's generalizability to unseen scenarios in the real world. For example, in COCO, many object categories have a much higher co-occurrence with men compared to women, which can bias a DNN's prediction in favor of men. Recent works have focused on task-specific training strategies to handle bias in such scenarios, but fixing the available data is often ignored. In this paper, we propose a novel and more generic solution to address the contextual bias in the datasets by selecting a subset of the samples, which is fair in terms of the co-occurrence with various classes for a protected attribute. We introduce a data repair algorithm using the coefficient of variation, which can curate fair and contextually balanced data for a protected class(es). This helps in training a fair model irrespective of the task, architecture or training methodology. Our proposed solution is simple, effective, and can even be used in an active learning setting where the data labels are not present or being generated incrementally. We demonstrate the effectiveness of our algorithm for the task of object detection and multi-label image classification across different datasets. Through a series of experiments, we validate that curating contextually fair data helps make model predictions fair by balancing the true positive rate for the protected class across groups without compromising on the model's overall performance.

READ FULL TEXT

page 1

page 7

research
04/27/2023

FLAC: Fairness-Aware Representation Learning by Suppressing Attribute-Class Associations

Bias in computer vision systems can perpetuate or even amplify discrimin...
research
03/24/2022

Repairing Group-Level Errors for DNNs Using Weighted Regularization

Deep Neural Networks (DNNs) have been widely used in software making dec...
research
09/22/2021

Contrastive Learning for Fair Representations

Trained classification models can unintentionally lead to biased represe...
research
01/09/2020

Don't Judge an Object by Its Context: Learning to Overcome Contextual Bias

Existing models often leverage co-occurrences between objects and their ...
research
09/18/2022

Through a fair looking-glass: mitigating bias in image datasets

With the recent growth in computer vision applications, the question of ...
research
02/11/2023

Pushing the Accuracy-Group Robustness Frontier with Introspective Self-play

Standard empirical risk minimization (ERM) training can produce deep neu...
research
10/31/2022

Improving Fairness in Image Classification via Sketching

Fairness is a fundamental requirement for trustworthy and human-centered...

Please sign up or login with your details

Forgot password? Click here to reset