DeepAI
Log In Sign Up

Your fairness may vary: Group fairness of pretrained language models in toxic text classification

08/03/2021
by   Ioana Baldini, et al.
ibm
0

We study the performance-fairness trade-off in more than a dozen fine-tuned LMs for toxic text classification. We empirically show that no blanket statement can be made with respect to the bias of large versus regular versus compressed models. Moreover, we find that focusing on fairness-agnostic performance metrics can lead to models with varied fairness characteristics.

READ FULL TEXT VIEW PDF

page 1

page 2

page 3

page 4

01/18/2018

Fine-tuned Language Models for Text Classification

Transfer learning has revolutionized computer vision, but existing appro...
10/13/2022

Equi-Tuning: Group Equivariant Fine-Tuning of Pretrained Models

We introduce equi-tuning, a novel fine-tuning method that transforms (po...
03/11/2019

Nuanced Metrics for Measuring Unintended Bias with Real Data for Text Classification

Unintended bias in Machine Learning can manifest as systemic differences...
12/14/2021

Measuring Fairness with Biased Rulers: A Survey on Quantifying Biases in Pretrained Language Models

An increasing awareness of biased patterns in natural language processin...
06/22/2022

Input-agnostic Certified Group Fairness via Gaussian Parameter Smoothing

Only recently, researchers attempt to provide classification algorithms ...
11/04/2020

Debiasing classifiers: is reality at variance with expectation?

Many methods for debiasing classifiers have been proposed, but their eff...

References