Empirical Decision Rules for Improving the Uncertainty Reporting of Small Sample System Usability Scale Scores

01/02/2021
by   Nicholas Clark, et al.
0

The System Usability Scale (SUS) is a short, survey-based approach used to determine the usability of a system from an end user perspective once a prototype is available for assessment. Individual scores are gathered using a 10-question survey with the survey results reported in terms of central tendency (sample mean) as an estimate of the system's usability (the SUS study score), and confidence intervals on the sample mean are used to communicate uncertainty levels associated with this point estimate. When the number of individuals surveyed is large, the SUS study scores and accompanying confidence intervals relying upon the central limit theorem for support are appropriate. However, when only a small number of users are surveyed, reliance on the central limit theorem falls short, resulting in confidence intervals that suffer from parameter bound violations and interval widths that confound mappings to adjective and other constructed scales. These shortcomings are especially pronounced when the underlying SUS score data is skewed, as it is in many instances. This paper introduces an empirically-based remedy for such small-sample circumstances, proposing a set of decision rules that leverage either an extended bias-corrected accelerated (BCa) bootstrap confidence interval or an empirical Bayesian credibility interval about the sample mean to restore and bolster subsequent confidence interval accuracy. Data from historical SUS assessments are used to highlight shortfalls in current practices and to demonstrate the improvements these alternate approaches offer while remaining statistically defensible. A freely available, online application is introduced and discussed that automates SUS analysis under these decision rules, thereby assisting usability practitioners in adopting the advocated approaches.

READ FULL TEXT

page 16

page 21

page 22

page 23

research
02/07/2022

Valid confidence intervals for μ, σ when there is only one observation available

Portnoy (2019) considered the problem of constructing an optimal confide...
research
03/11/2017

The Accuracy of Confidence Intervals for Field Normalised Indicators

When comparing the average citation impact of research groups, universit...
research
06/06/2018

Error analysis for small-sample, high-variance data: Cautions for bootstrapping and Bayesian bootstrapping

Recent advances in molecular simulations allow the direct evaluation of ...
research
03/29/2023

Did You Mean...? Confidence-based Trade-offs in Semantic Parsing

We illustrate how a calibrated model can help balance common trade-offs ...
research
01/30/2015

Confidence intervals for AB-test

AB-testing is a very popular technique in web companies since it makes i...
research
11/27/2022

An Empirical Bayes Approach for Constructing the Confidence Intervals of Clonality and Entropy

This paper is motivated by the need to quantify human immune responses t...
research
08/18/2023

Generative Machine Listener

We show how a neural network can be trained on individual intrusive list...

Please sign up or login with your details

Forgot password? Click here to reset