Targeted Data Augmentation for bias mitigation

The development of fair and ethical AI systems requires careful consideration of bias mitigation, an area often overlooked or ignored. In this study, we introduce a novel and efficient approach for addressing biases called Targeted Data Augmentation (TDA), which leverages classical data augmentation techniques to tackle the pressing issue of bias in data and models. Unlike the laborious task of removing biases, our method proposes to insert biases instead, resulting in improved performance. To identify biases, we annotated two diverse datasets: a dataset of clinical skin lesions and a dataset of male and female faces. These bias annotations are published for the first time in this study, providing a valuable resource for future research. Through Counterfactual Bias Insertion, we discovered that biases associated with the frame, ruler, and glasses had a significant impact on models. By randomly introducing biases during training, we mitigated these biases and achieved a substantial decrease in bias measures, ranging from two-fold to more than 50-fold, while maintaining a negligible increase in the error rate.

READ FULL TEXT
research
08/18/2023

Data augmentation and explainability for bias discovery and mitigation in deep learning

This dissertation explores the impact of bias in deep neural networks an...
research
05/26/2023

Leveraging Domain Knowledge for Inclusive and Bias-aware Humanitarian Response Entry Classification

Accurate and rapid situation analysis during humanitarian crises is crit...
research
05/25/2022

Does Your Model Classify Entities Reasonably? Diagnosing and Mitigating Spurious Correlations in Entity Typing

The entity typing task aims at predicting one or more words or phrases t...
research
06/30/2021

Fair Visual Recognition in Limited Data Regime using Self-Supervision and Self-Distillation

Deep learning models generally learn the biases present in the training ...
research
05/06/2023

Algorithmic Bias, Generalist Models,and Clinical Medicine

The technical landscape of clinical machine learning is shifting in ways...
research
08/03/2023

NBIAS: A Natural Language Processing Framework for Bias Identification in Text

Bias in textual data can lead to skewed interpretations and outcomes whe...
research
06/30/2022

Personalized Detection of Cognitive Biases in Actions of Users from Their Logs: Anchoring and Recency Biases

Cognitive biases are mental shortcuts humans use in dealing with informa...

Please sign up or login with your details

Forgot password? Click here to reset