Bias at a Second Glance: A Deep Dive into Bias for German Educational Peer-Review Data Modeling

09/21/2022
by   Thiemo Wambsganss, et al.
0

Natural Language Processing (NLP) has become increasingly utilized to provide adaptivity in educational applications. However, recent research has highlighted a variety of biases in pre-trained language models. While existing studies investigate bias in different domains, they are limited in addressing fine-grained analysis on educational and multilingual corpora. In this work, we analyze bias across text and through multiple architectures on a corpus of 9,165 German peer-reviews collected from university students over five years. Notably, our corpus includes labels such as helpfulness, quality, and critical aspect ratings from the peer-review recipient as well as demographic attributes. We conduct a Word Embedding Association Test (WEAT) analysis on (1) our collected corpus in connection with the clustered labels, (2) the most common pre-trained German language models (T5, BERT, and GPT-2) and GloVe embeddings, and (3) the language models after fine-tuning on our collected data-set. In contrast to our initial expectations, we found that our collected corpus does not reveal many biases in the co-occurrence analysis or in the GloVe embeddings. However, the pre-trained German language models find substantial conceptual, racial, and gender bias and have significant changes in bias across conceptual and racial axes during fine-tuning on the peer-review data. With our research, we aim to contribute to the fourth UN sustainability goal (quality education) with a novel dataset, an understanding of biases in natural language education data, and the potential harms of not counteracting biases in language models for educational tasks.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/21/2022

Towards an Enhanced Understanding of Bias in Pre-trained Neural Language Models: A Survey with Special Emphasis on Affective Bias

The remarkable progress in Natural Language Processing (NLP) brought abo...
research
05/03/2021

Impact of Gender Debiased Word Embeddings in Language Modeling

Gender, race and social biases have recently been detected as evident ex...
research
07/31/2018

Gender Bias in Neural Natural Language Processing

We examine whether neural natural language processing (NLP) systems refl...
research
10/27/2020

Unmasking Contextual Stereotypes: Measuring and Mitigating BERT's Gender Bias

Contextualized word embeddings have been replacing standard embeddings a...
research
11/07/2022

Investigating Fairness Disparities in Peer Review: A Language Model Enhanced Approach

Double-blind peer review mechanism has become the skeleton of academic r...
research
01/21/2023

Blacks is to Anger as Whites is to Joy? Understanding Latent Affective Bias in Large Pre-trained Neural Language Models

Groundbreaking inventions and highly significant performance improvement...
research
11/15/2022

Mind Your Bias: A Critical Review of Bias Detection Methods for Contextual Language Models

The awareness and mitigation of biases are of fundamental importance for...

Please sign up or login with your details

Forgot password? Click here to reset