contextual: Evaluating Contextual Multi-Armed Bandit Problems in R

11/06/2018
by   Robin van Emden, et al.
0

Over the past decade, contextual bandit algorithms have been gaining in popularity due to their effectiveness and flexibility in solving sequential decision problems---from online advertising and finance to clinical trial design and personalized medicine. At the same time, there are, as of yet, surprisingly few options that enable researchers and practitioners to simulate and compare the wealth of new and existing bandit algorithms in a standardized way. To help close this gap between analytical research and empirical evaluation the current paper introduces the object-oriented R package "contextual": a user-friendly and, through its object-oriented structure, easily extensible framework that facilitates parallelized comparison of contextual and context-free bandit policies through both simulation and offline analysis.

READ FULL TEXT
research
06/26/2020

Online learning with Corrupted context: Corrupted Contextual Bandits

We consider a novel variant of the contextual bandit problem (i.e., the ...
research
07/13/2020

Contextual Bandit with Missing Rewards

We consider a novel variant of the contextual bandit problem (i.e., the ...
research
05/10/2017

Context Attentive Bandits: Contextual Bandit with Restricted Context

We consider a novel formulation of the multi-armed bandit model, which w...
research
02/04/2022

Tsetlin Machine for Solving Contextual Bandit Problems

This paper introduces an interpretable contextual bandit algorithm using...
research
06/24/2023

Contextual aggregation and rapid updating of trial outcomes within a user-friendly open-source environment

The delayed and incomplete availability of historical findings and the l...
research
08/21/2019

Exploring Offline Policy Evaluation for the Continuous-Armed Bandit Problem

The (contextual) multi-armed bandit problem (MAB) provides a formalizati...
research
08/08/2023

AdaptEx: A Self-Service Contextual Bandit Platform

This paper presents AdaptEx, a self-service contextual bandit platform w...

Please sign up or login with your details

Forgot password? Click here to reset