A Biologically Plausible Benchmark for Contextual Bandit Algorithms in Precision Oncology Using in vitro Data

11/11/2019
by   Niklas T. Rindtorff, et al.
0

Precision oncology, the genetic sequencing of tumors to identify druggable targets, has emerged as the standard of care in the treatment of many cancers. Nonetheless, due to the pace of therapy development and variability in patient information, designing effective protocols for individual treatment assignment in a sample-efficient way remains a major challenge. One promising approach to this problem is to frame precision oncology treatment as a contextual bandit problem and to apply sequential decision-making algorithms designed to minimize regret in this setting. However, a clear prerequisite for considering this methodology in high-stakes clinical decisions is careful benchmarking to understand realistic costs and benefits. Here, we propose a benchmark dataset to evaluate contextual bandit algorithms based on real in vitro drug response of approximately 900 cancer cell lines. Specifically, we curated a dataset of complete treatment responses for a subset of 7 treatments from prior in vitro studies. This allows us to compute the regret of proposed decision policies using biologically plausible counterfactuals. We ran a suite of Bayesian bandit algorithms on our benchmark, and found that the methods accumulate less regret over a sequence of treatment assignment tasks than a rule-based baseline derived from current clinical practice. This effect was more pronounced when genomic information was included as context. We expect this work to be a starting point for evaluation of both the unique structural requirements and ethical implications for real-world testing of bandit based clinical decision support.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/05/2022

A Deep Bayesian Bandits Approach for Anticancer Therapy: Exploration via Functional Prior

Learning personalized cancer treatment with machine learning holds great...
research
09/01/2018

A Contextual-bandit-based Approach for Informed Decision-making in Clinical Trials

Clinical trials involving multiple treatments utilize randomization of t...
research
05/09/2022

Selectively Contextual Bandits

Contextual bandits are widely used in industrial personalization systems...
research
08/10/2021

Bandit Algorithms for Precision Medicine

The Oxford English Dictionary defines precision medicine as "medical car...
research
07/02/2020

Learning Individualized Treatment Rules with Estimated Translated Inverse Propensity Score

Randomized controlled trials typically analyze the effectiveness of trea...
research
05/29/2023

Contextual Bandits with Budgeted Information Reveal

Contextual bandit algorithms are commonly used in digital health to reco...
research
05/06/2020

DTR Bandit: Learning to Make Response-Adaptive Decisions With Low Regret

Dynamic treatment regimes (DTRs) for are personalized, sequential treatm...

Please sign up or login with your details

Forgot password? Click here to reset