Novel Techniques to Assess Predictive Systems and Reduce Their Alarm Burden

02/10/2021
by   Jonathan A. Handler, et al.
0

The performance of a binary classifier ("predictor") depends heavily upon the context ("workflow") in which it operates. Classic measures of predictor performance do not reflect the realized utility of predictors unless certain implied workflow assumptions are met. Failure to meet these implied assumptions results in suboptimal classifier implementations and a mismatch between predicted or assessed performance and the actual performance obtained in real-world deployments. The mismatch commonly arises when multiple predictions can be made for the same event, the event is relatively rare, and redundant true positive predictions for the same event add little value, e.g., a system that makes a prediction each minute, repeatedly issuing interruptive alarms for a predicted event that may never occur. We explain why classic metrics do not correctly represent the performance of predictors in such contexts, and introduce an improved performance assessment technique ("u-metrics") using utility functions to score each prediction. U-metrics explicitly account for variability in prediction utility arising from temporal relationships. Compared to traditional performance measures, u-metrics more accurately reflect the real-world benefits and costs of a predictor operating in a workflow context. The difference can be significant. We also describe the use of "snoozing," a method whereby predictions are suppressed for a period of time, commonly improving predictor performance by reducing false positives while retaining the capture of events. Snoozing is especially useful when predictors generate interruptive alerts, as so often happens in clinical practice. Utility-based performance metrics correctly predict and track the performance benefits of snoozing, whereas traditional performance metrics do not.

READ FULL TEXT
research
10/22/2022

Algorithms with Prediction Portfolios

The research area of algorithms with predictions has seen recent success...
research
10/19/2022

On Learning Fairness and Accuracy on Multiple Subgroups

We propose an analysis in fair learning that preserves the utility of th...
research
05/24/2021

Optimized conformal classification using gradient descent approximation

Conformal predictors are an important class of algorithms that allow pre...
research
07/05/2017

The impossibility of "fairness": a generalized impossibility result for decisions

Various measures can be used to estimate bias or unfairness in a predict...
research
04/22/2019

Tracking and Improving Information in the Service of Fairness

As algorithmic prediction systems have become widespread, fears that the...
research
10/12/2022

DiPA: Diverse and Probabilistically Accurate Interactive Prediction

Accurate prediction is important for operating an autonomous vehicle in ...
research
06/02/2020

Performance metrics for intervention-triggering prediction models do not reflect an expected reduction in outcomes from using the model

Clinical researchers often select among and evaluate risk prediction mod...

Please sign up or login with your details

Forgot password? Click here to reset