Multi-annotator Deep Learning: A Probabilistic Framework for Classification

04/05/2023
by   Marek Herde, et al.
1

Solving complex classification tasks using deep neural networks typically requires large amounts of annotated data. However, corresponding class labels are noisy when provided by error-prone annotators, e.g., crowd workers. Training standard deep neural networks leads to subpar performances in such multi-annotator supervised learning settings. We address this issue by presenting a probabilistic training framework named multi-annotator deep learning (MaDL). A ground truth and an annotator performance model are jointly trained in an end-to-end learning approach. The ground truth model learns to predict instances' true class labels, while the annotator performance model infers probabilistic estimates of annotators' performances. A modular network architecture enables us to make varying assumptions regarding annotators' performances, e.g., an optional class or instance dependency. Further, we learn annotator embeddings to estimate annotators' densities within a latent space as proxies of their potentially correlated annotations. Together with a weighted loss function, we improve the learning from correlated annotation patterns. In a comprehensive evaluation, we examine three research questions about multi-annotator supervised learning. Our findings indicate MaDL's state-of-the-art performance and robustness against many correlated, spamming annotators.

READ FULL TEXT

page 18

page 24

research
07/11/2021

Learning from Crowds with Sparse and Imbalanced Annotations

Traditional supervised learning requires ground truth labels for the tra...
research
04/07/2020

Learning from Imperfect Annotations

Many machine learning systems today are trained on large amounts of huma...
research
10/22/2020

On the Power of Deep but Naive Partial Label Learning

Partial label learning (PLL) is a class of weakly supervised learning wh...
research
08/21/2023

Label Selection Approach to Learning from Crowds

Supervised learning, especially supervised deep learning, requires large...
research
05/15/2017

Learning Probabilistic Programs Using Backpropagation

Probabilistic modeling enables combining domain knowledge with learning ...
research
07/23/2018

Deep Learning from Label Proportions for Emphysema Quantification

We propose an end-to-end deep learning method that learns to estimate em...
research
08/05/2019

Mass Estimation from Images using Deep Neural Network and Sparse Ground Truth

Supervised learning is the workhorse for regression and classification t...

Please sign up or login with your details

Forgot password? Click here to reset