Overall Agreement for Multiple Raters with Replicated Measurements

06/07/2020
by   Tongrong Wang, et al.
0

Multiple raters are often needed to be used interchangeably in practice for measurement or evaluation. Assessing agreement among these multiple raters via agreement indices are necessary before their participation. While the intuitively appealing agreement indices such as coverage probability and total deviation index, and relative area under coverage probability curve, have been extended for assessing overall agreement among multiple raters, these extensions have limitations. The existing overall agreement indices either require normality and homogeneity assumptions or did not preserve the intuitive interpretation of the indices originally defined for two raters. In this paper, we propose a new set of overall agreement indices based on maximum pairwise differences among all raters. The proposed new overall coverage probability, overall total deviation index and relative area under overall coverage probability curve retain the original intuitive interpretation from the pairwise version. Without making any distributional assumption, we also propose a new unified nonparametric estimation and inference approach for the overall indices based on generalized estimating equations that can accommodate replications made by the same rater. Under mild assumptions, the proposed variance estimator is shown to achieve efficiency bound under independent working correlation matrix. Simulation studies under different scenarios are conducted to assess the performance of the proposed estimation and inference approach with and without replications. We illustrate the methodology by using a blood pressure data with three raters who made three replications on each subjects.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/10/2018

Axiomatizations of inconsistency indices for triads

Pairwise comparison matrices often exhibit inconsistency, therefore, a n...
research
01/07/2019

Understanding partition comparison indices based on counting object pairs

In unsupervised machine learning, agreement between partitions is common...
research
05/03/2018

Sound Colless-like balance indices for multifurcating trees

The Colless index is one of the most popular and natural balance indices...
research
08/06/2023

Empirical Optimal Risk to Quantify Model Trustworthiness for Failure Detection

Failure detection (FD) in AI systems is a crucial safeguard for the depl...
research
08/12/2022

Sparse Probability of Agreement

Measuring inter-annotator agreement is important for annotation tasks, b...
research
09/25/2018

Assessing Method Agreement for Paired Repeated Binary Measurements

Method comparison studies are essential for development in medical and c...
research
07/25/2015

Truth Serums for Massively Crowdsourced Evaluation Tasks

A major challenge in crowdsourcing evaluation tasks like labeling object...

Please sign up or login with your details

Forgot password? Click here to reset