Survey Bandits with Regret Guarantees

We consider a variant of the contextual bandit problem. In standard contextual bandits, when a user arrives we get the user's complete feature vector and then assign a treatment (arm) to that user. In a number of applications (like healthcare), collecting features from users can be costly. To address this issue, we propose algorithms that avoid needless feature collection while maintaining strong regret guarantees.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/05/2023

Proportional Response: Contextual Bandits for Simple and Cumulative Regret Minimization

Simple regret minimization is a critical problem in learning optimal tre...
research
05/09/2022

Selectively Contextual Bandits

Contextual bandits are widely used in industrial personalization systems...
research
02/07/2023

Leveraging User-Triggered Supervision in Contextual Bandits

We study contextual bandit (CB) problems, where the user can sometimes r...
research
06/21/2020

An Opportunistic Bandit Approach for User Interface Experimentation

Facing growing competition from online rivals, the retail industry is in...
research
12/06/2021

Contextual Bandit Applications in Customer Support Bot

Virtual support agents have grown in popularity as a way for businesses ...
research
06/09/2021

Contextual Recommendations and Low-Regret Cutting-Plane Algorithms

We consider the following variant of contextual linear bandits motivated...
research
10/24/2022

Conditionally Risk-Averse Contextual Bandits

We desire to apply contextual bandits to scenarios where average-case st...

Please sign up or login with your details

Forgot password? Click here to reset