Type B Reflexivization as an Unambiguous Testbed for Multilingual Multi-Task Gender Bias

09/24/2020
by   Ana Valeria Gonzalez, et al.
0

The one-sided focus on English in previous studies of gender bias in NLP misses out on opportunities in other languages: English challenge datasets such as GAP and WinoGender highlight model preferences that are "hallucinatory", e.g., disambiguating gender-ambiguous occurrences of 'doctor' as male doctors. We show that for languages with type B reflexivization, e.g., Swedish and Russian, we can construct multi-task challenge datasets for detecting gender bias that lead to unambiguously wrong model predictions: In these languages, the direct translation of 'the doctor removed his mask' is not ambiguous between a coreferential reading and a disjoint reading. Instead, the coreferential reading requires a non-gendered pronoun, and the gendered, possessive pronouns are anti-reflexive. We present a multilingual, multi-task challenge dataset, which spans four languages and four NLP tasks and focuses only on this phenomenon. We find evidence for gender bias across all task-language combinations and correlate model bias with national labor market statistics.

READ FULL TEXT
research
10/12/2020

Gender Coreference and Bias Evaluation at WMT 2020

Gender bias in machine translation can manifest when choosing gender inf...
research
11/22/2021

Investigating Cross-Linguistic Gender Bias in Hindi-English Across Domains

Measuring, evaluating and reducing Gender Bias has come to the forefront...
research
05/12/2022

Mitigating Gender Stereotypes in Hindi and Marathi

As the use of natural language processing increases in our day-to-day li...
research
06/03/2019

Gendered Ambiguous Pronouns Shared Task: Boosting Model Confidence by Evidence Pooling

This paper presents a strong set of results for resolving gendered ambig...
research
06/02/2020

A Multi-Task Comparator Framework for Kinship Verification

Approaches for kinship verification often rely on cosine distances betwe...
research
10/18/2021

The Arabic Parallel Gender Corpus 2.0: Extensions and Analyses

Gender bias in natural language processing (NLP) applications, particula...
research
06/20/2022

Fewer Errors, but More Stereotypes? The Effect of Model Size on Gender Bias

The size of pretrained models is increasing, and so is their performance...

Please sign up or login with your details

Forgot password? Click here to reset