Consistency is Key: Disentangling Label Variation in Natural Language Processing with Intra-Annotator Agreement

01/25/2023
by   Gavin Abercrombie, et al.
0

We commonly use agreement measures to assess the utility of judgements made by human annotators in Natural Language Processing (NLP) tasks. While inter-annotator agreement is frequently used as an indication of label reliability by measuring consistency between annotators, we argue for the additional use of intra-annotator agreement to measure label stability over time. However, in a systematic review, we find that the latter is rarely reported in this field. Calculating these measures can act as important quality control and provide insights into why annotators disagree. We propose exploratory annotation experiments to investigate the relationships between these measures and perceptions of subjectivity and ambiguity in text items.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/24/2023

Understanding and Predicting Human Label Variation in Natural Language Inference through Explanation

Human label variation (Plank 2022), or annotation disagreement, exists i...
research
06/26/2023

Inter-Annotator Agreement in the Wild: Uncovering Its Emerging Roles and Considerations in Real-World Scenarios

Inter-Annotator Agreement (IAA) is commonly used as a measure of label c...
research
12/15/2022

Measuring Annotator Agreement Generally across Complex Structured, Multi-object, and Free-text Annotation Tasks

When annotators label data, a key metric for quality assurance is inter-...
research
09/29/2020

Aligning Intraobserver Agreement by Transitivity

Annotation reproducibility and accuracy rely on good consistency within ...
research
10/28/2022

"It's Not Just Hate”: A Multi-Dimensional Perspective on Detecting Harmful Speech Online

Well-annotated data is a prerequisite for good Natural Language Processi...
research
09/17/2022

DiPietro-Hazari Kappa: A Novel Metric for Assessing Labeling Quality via Annotation

Data is a key component of modern machine learning, but statistics for a...
research
03/07/2018

Sklar's Omega: A Gaussian Copula-Based Framework for Assessing Agreement

The statistical measurement of agreement is important in a number of fie...

Please sign up or login with your details

Forgot password? Click here to reset