Context Attentive Bandits: Contextual Bandit with Restricted Context

05/10/2017
by   Djallel Bouneffouf, et al.
0

We consider a novel formulation of the multi-armed bandit model, which we call the contextual bandit with restricted context, where only a limited number of features can be accessed by the learner at every iteration. This novel formulation is motivated by different online problems arising in clinical trials, recommender systems and attention modeling. Herein, we adapt the standard multi-armed bandit algorithm known as Thompson Sampling to take advantage of our restricted context setting, and propose two novel algorithms, called the Thompson Sampling with Restricted Context(TSRC) and the Windows Thompson Sampling with Restricted Context(WTSRC), for handling stationary and nonstationary environments, respectively. Our empirical results demonstrate advantages of the proposed approaches on several real-life datasets

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/26/2020

Online learning with Corrupted context: Corrupted Contextual Bandits

We consider a novel variant of the contextual bandit problem (i.e., the ...
research
10/15/2020

Double-Linear Thompson Sampling for Context-Attentive Bandits

In this paper, we analyze and extend an online learning framework known ...
research
01/04/2021

Etat de l'art sur l'application des bandits multi-bras

The Multi-armed bandit offer the advantage to learn and exploit the alre...
research
11/22/2017

Customized Nonlinear Bandits for Online Response Selection in Neural Conversation Models

Dialog response selection is an important step towards natural response ...
research
10/23/2019

Diversifying Database Activity Monitoring with Bandits

Database activity monitoring (DAM) systems are commonly used by organiza...
research
07/15/2023

On the Robustness of Epoch-Greedy in Multi-Agent Contextual Bandit Mechanisms

Efficient learning in multi-armed bandit mechanisms such as pay-per-clic...
research
11/06/2018

contextual: Evaluating Contextual Multi-Armed Bandit Problems in R

Over the past decade, contextual bandit algorithms have been gaining in ...

Please sign up or login with your details

Forgot password? Click here to reset