Data Augmentation via Subgroup Mixup for Improving Fairness

09/13/2023
by   Madeline Navarro, et al.
0

In this work, we propose data augmentation via pairwise mixup across subgroups to improve group fairness. Many real-world applications of machine learning systems exhibit biases across certain groups due to under-representation or training data that reflects societal biases. Inspired by the successes of mixup for improving classification performance, we develop a pairwise mixup scheme to augment training data and encourage fair and accurate decision boundaries for all subgroups. Data augmentation for group fairness allows us to add new samples of underrepresented groups to balance subpopulations. Furthermore, our method allows us to use the generalization ability of mixup to improve both fairness and accuracy. We compare our proposed mixup to existing data augmentation and bias mitigation approaches on both synthetic simulations and real-world benchmark fair classification data, demonstrating that we are able to achieve fair outcomes with robust if not improved accuracy.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/11/2021

Fair Mixup: Fairness via Interpolation

Training classifiers under fairness constraints such as group fairness, ...
research
07/27/2020

Towards Accuracy-Fairness Paradox: Adversarial Example-based Data Augmentation for Visual Debiasing

Machine learning fairness concerns about the biases towards certain prot...
research
10/08/2021

Distinguishing rule- and exemplar-based generalization in learning systems

Despite the increasing scale of datasets in machine learning, generaliza...
research
01/17/2022

Fair Group-Shared Representations with Normalizing Flows

The issue of fairness in machine learning stems from the fact that histo...
research
11/24/2021

Fairness for AUC via Feature Augmentation

We study fairness in the context of classification where the performance...
research
06/02/2023

Affinity Clustering Framework for Data Debiasing Using Pairwise Distribution Discrepancy

Group imbalance, resulting from inadequate or unrepresentative data coll...
research
05/23/2019

Generative Adversarial Networks for Mitigating Biases in Machine Learning Systems

In this paper, we propose a new framework for mitigating biases in machi...

Please sign up or login with your details

Forgot password? Click here to reset