Towards Equal Opportunity Fairness through Adversarial Learning

03/12/2022
by   Xudong Han, et al.
0

Adversarial training is a common approach for bias mitigation in natural language processing. Although most work on debiasing is motivated by equal opportunity, it is not explicitly captured in standard adversarial training. In this paper, we propose an augmented discriminator for adversarial training, which takes the target class as input to create richer features and more explicitly model equal opportunity. Experimental results over two datasets show that our method substantially improves over standard adversarial debiasing methods, in terms of the performance–fairness trade-off.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/05/2022

Optimising Equal Opportunity Fairness in Model Training

Real-world datasets often encode stereotypes and societal biases. Such b...
research
01/25/2021

Diverse Adversaries for Mitigating Bias in Training

Adversarial learning can learn fairer and less biased models of language...
research
04/12/2021

Targeted Adversarial Training for Natural Language Understanding

We present a simple yet effective Targeted Adversarial Training (TAT) al...
research
05/14/2023

Algorithmic Pluralism: A Structural Approach Towards Equal Opportunity

While the idea of equal opportunity enjoys a broad consensus, many disag...
research
04/11/2021

Adversarial Training as Stackelberg Game: An Unrolled Optimization Approach

Adversarial training has been shown to improve the generalization perfor...
research
07/28/2021

Imbalanced Adversarial Training with Reweighting

Adversarial training has been empirically proven to be one of the most e...
research
11/28/2017

End-to-end Adversarial Learning for Generative Conversational Agents

This paper presents a new adversarial learning method for generative con...

Please sign up or login with your details

Forgot password? Click here to reset