On Testing for Biases in Peer Review

12/31/2019
by   Ivan Stelmakh, et al.
0

We consider the issue of biases in scholarly research, specifically, in peer review. There is a long standing debate on whether exposing author identities to reviewers induces biases against certain groups, and our focus is on designing tests to detect the presence of such biases. Our starting point is a remarkable recent work by Tomkins, Zhang and Heavlin which conducted a controlled, large-scale experiment to investigate existence of biases in the peer reviewing of the WSDM conference. We present two sets of results in this paper. The first set of results is negative, and pertains to the statistical tests and the experimental setup used in the work of Tomkins et al. We show that the test employed therein does not guarantee control over false alarm probability and under correlations between relevant variables coupled with any of the following conditions, with high probability, can declare a presence of bias when it is in fact absent: (a) measurement error, (b) model mismatch, (c) reviewer calibration. Moreover, we show that the setup of their experiment may itself inflate false alarm probability if (d) bidding is performed in non-blind manner or (e) popular reviewer assignment procedure is employed. Our second set of results is positive and is built around a novel approach to testing for biases that we propose. We present a general framework for testing for biases in (single vs. double blind) peer review. We then design hypothesis tests that under minimal assumptions guarantee control over false alarm probability and non-trivial power even under conditions (a)–(c) as well as propose an alternative experimental setup which mitigates issues (d) and (e). Finally, we show that no statistical test can improve over the non-parametric tests we consider in terms of the assumptions required to control for the false alarm probability.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/04/2018

Case for the double-blind peer review

Peer review is a process designed to produce a fair assessment of resear...
research
11/30/2020

A Large Scale Randomized Controlled Trial on Herding in Peer-Review Discussions

Peer review is the backbone of academia and humans constitute a cornerst...
research
10/29/2020

Uncovering Latent Biases in Text: Method and Application to Peer Review

Quantifying systematic disparities in numerical quantities such as emplo...
research
11/07/2022

Investigating Fairness Disparities in Peer Review: A Language Model Enhanced Approach

Double-blind peer review mechanism has become the skeleton of academic r...
research
10/08/2020

Catch Me if I Can: Detecting Strategic Behaviour in Peer Assessment

We consider the issue of strategic behaviour in various peer-assessment ...
research
10/22/2020

Drift Detection in Episodic Data: Detect When Your Agent Starts Faltering

Detection of deterioration of agent performance in dynamic environments ...

Please sign up or login with your details

Forgot password? Click here to reset