Challenges and strategies for running controlled crowdsourcing experiments

11/05/2020
by   Jorge Ramirez, et al.
0

This paper reports on the challenges and lessons we learned while running controlled experiments in crowdsourcing platforms. Crowdsourcing is becoming an attractive technique to engage a diverse and large pool of subjects in experimental research, allowing researchers to achieve levels of scale and completion times that would otherwise not be feasible in lab settings. However, the scale and flexibility comes at the cost of multiple and sometimes unknown sources of bias and confounding factors that arise from technical limitations of crowdsourcing platforms and from the challenges of running controlled experiments in the "wild". In this paper, we take our experience in running systematic evaluations of task design as a motivating example to explore, describe, and quantify the potential impact of running uncontrolled crowdsourcing experiments and derive possible coping strategies. Among the challenges identified, we can mention sampling bias, controlling the assignment of subjects to experimental conditions, learning effects, and reliability of crowdsourcing results. According to our empirical studies, the impact of potential biases and confounding factors can amount to a 38% loss in the utility of the data collected in uncontrolled settings; and it can significantly change the outcome of experiments. These issues ultimately inspired us to implement CrowdHub, a system that sits on top of major crowdsourcing platforms and allows researchers and practitioners to run controlled crowdsourcing projects.

READ FULL TEXT
research
09/06/2019

CrowdHub: Extending crowdsourcing platforms for the controlled evaluation of tasks designs

We present CrowdHub, a tool for running systematic evaluations of task d...
research
07/28/2021

On the state of reporting in crowdsourcing experiments and a checklist to aid current practices

Crowdsourcing is being increasingly adopted as a platform to run studies...
research
09/22/2020

Can we trust online crowdworkers? Comparing online and offline participants in a preference test of virtual agents

Conducting user studies is a crucial component in many scientific fields...
research
03/25/2020

Impact of the Number of Votes on the Reliability and Validity of Subjective Speech Quality Assessment in the Crowdsourcing Approach

The subjective quality of transmitted speech is traditionally assessed i...
research
05/17/2019

Comparison-Based Framework for Psychophysics: Lab versus Crowdsourcing

Traditionally, psychophysical experiments are conducted by repeated meas...
research
03/15/2021

Considerations and Pitfalls in Controlled Experiments on Code Comprehension

Understanding program code is a complicated endeavor. As such, myriad di...
research
08/25/2022

Lessons Learned from a Bare-metal Evaluation of Erasure Coding Algorithms in P2P Networks

We have built a bare-metal testbed in order to perform large-scale, repr...

Please sign up or login with your details

Forgot password? Click here to reset