Crowd Labeling: a survey

01/13/2013
by   Jafar Muhammadi, et al.
0

Recently, there has been a burst in the number of research projects on human computation via crowdsourcing. Multiple choice (or labeling) questions could be referred to as a common type of problem which is solved by this approach. As an application, crowd labeling is applied to find true labels for large machine learning datasets. Since crowds are not necessarily experts, the labels they provide are rather noisy and erroneous. This challenge is usually resolved by collecting multiple labels for each sample, and then aggregating them to estimate the true label. Although the mechanism leads to high-quality labels, it is not actually cost-effective. As a result, efforts are currently made to maximize the accuracy in estimating true labels, while fixing the number of acquired labels. This paper surveys methods to aggregate redundant crowd labels in order to estimate unknown true labels. It presents a unified statistical latent model where the differences among popular methods in the field correspond to different choices for the parameters of the model. Afterwards, algorithms to make inference on these models will be surveyed. Moreover, adaptive methods which iteratively collect labels based on the previously collected labels and estimated models will be discussed. In addition, this paper compares the distinguished methods, and provides guidelines for future work required to address the current open issues.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/23/2018

Optimizing the Wisdom of the Crowd: Inference, Learning, and Teaching

The unprecedented demand for large amount of data has catalyzed the tren...
research
11/19/2022

A Light-weight, Effective and Efficient Model for Label Aggregation in Crowdsourcing

Due to the noises in crowdsourced labels, label aggregation (LA) has eme...
research
12/13/2017

Ballpark Crowdsourcing: The Wisdom of Rough Group Comparisons

Crowdsourcing has become a popular method for collecting labeled trainin...
research
04/26/2018

Weak Labeling for Crowd Learning

Crowdsourcing has become very popular among the machine learning communi...
research
12/05/2012

Evaluating Classifiers Without Expert Labels

This paper considers the challenge of evaluating a set of classifiers, a...
research
06/24/2022

How many labelers do you have? A closer look at gold-standard labels

The construction of most supervised learning datasets revolves around co...
research
06/20/2019

Latent Distribution Assumption for Unbiased and Consistent Consensus Modelling

We study the problem of aggregation noisy labels. Usually, it is solved ...

Please sign up or login with your details

Forgot password? Click here to reset