Training Fair Deep Neural Networks by Balancing Influence

01/15/2022
by   Haonan Wang, et al.
0

Most fair machine learning methods either highly rely on the sensitive information of the training samples or require a large modification on the target models, which hinders their practical application. To address this issue, we propose a two-stage training algorithm named FAIRIF. It minimizes the loss over the reweighted data set (second stage) where the sample weights are computed to balance the model performance across different demographic groups (first stage). FAIRIF can be applied on a wide range of models trained by stochastic gradient descent without changing the model, while only requiring group annotations on a small validation set to compute sample weights. Theoretically, we show that, in the classification setting, three notions of disparity among different groups can be mitigated by training with the weights. Experiments on synthetic data sets demonstrate that FAIRIF yields models with better fairness-utility trade-offs against various types of bias; and on real-world data sets, we show the effectiveness and scalability of FAIRIF. Moreover, as evidenced by the experiments with pretrained models, FAIRIF is able to alleviate the unfairness issue of pretrained models without hurting their performance.

READ FULL TEXT
research
06/03/2021

BiFair: Training Fair Models with Bilevel Optimization

Prior studies have shown that, training machine learning models via empi...
research
09/17/2019

A Distributed Fair Machine Learning Framework with Private Demographic Data Protection

Fair machine learning has become a significant research topic with broad...
research
06/07/2023

Migrate Demographic Group For Fair GNNs

Graph Neural networks (GNNs) have been applied in many scenarios due to ...
research
08/28/2023

Fair Few-shot Learning with Auxiliary Sets

Recently, there has been a growing interest in developing machine learni...
research
05/17/2022

Perturbation of Deep Autoencoder Weights for Model Compression and Classification of Tabular Data

Fully connected deep neural networks (DNN) often include redundant weigh...
research
03/01/2021

Adaptive Sampling for Minimax Fair Classification

Machine learning models trained on imbalanced datasets can often end up ...
research
06/11/2020

Adaptive Sampling to Reduce Disparate Performance

Existing methods for reducing disparate performance of a classifier acro...

Please sign up or login with your details

Forgot password? Click here to reset