TrueLabel + Confusions: A Spectrum of Probabilistic Models in Analyzing Multiple Ratings

06/18/2012
by   Chao Liu, et al.
0

This paper revisits the problem of analyzing multiple ratings given by different judges. Different from previous work that focuses on distilling the true labels from noisy crowdsourcing ratings, we emphasize gaining diagnostic insights into our in-house well-trained judges. We generalize the well-known DawidSkene model (Dawid & Skene, 1979) to a spectrum of probabilistic models under the same "TrueLabel + Confusion" paradigm, and show that our proposed hierarchical Bayesian model, called HybridConfusion, consistently outperforms DawidSkene on both synthetic and real-world data sets.

READ FULL TEXT

page 5

page 8

research
05/15/2017

Quantifying Aspect Bias in Ordinal Ratings using a Bayesian Approach

User opinions expressed in the form of ratings can influence an individu...
research
10/19/2020

Rater: An R Package for Fitting Statistical Models of Repeated Categorical Ratings

A common occurrence in many disciplines is the need to assign a set of i...
research
08/31/2019

Humor Detection: A Transformer Gets the Last Laugh

Much previous work has been done in attempting to identify humor in text...
research
11/21/2019

Reinforcing an Image Caption Generator Using Off-Line Human Feedback

Human ratings are currently the most accurate way to assess the quality ...
research
06/27/2012

Visualization of Collaborative Data

Collaborative data consist of ratings relating two distinct sets of obje...
research
08/08/2022

Reliability of Solutions in Linear Ordering Problem: New Probabilistic Insight and Algorithms

In this work, our goal is to characterize the reliability of the solutio...
research
10/13/2022

Utilizing supervised models to infer consensus labels and their quality from data with multiple annotators

Real-world data for classification is often labeled by multiple annotato...

Please sign up or login with your details

Forgot password? Click here to reset