Etat de l'art sur l'application des bandits multi-bras

01/04/2021
by   Djallel Bouneffouf, et al.
0

The Multi-armed bandit offer the advantage to learn and exploit the already learnt knowledge at the same time. This capability allows this approach to be applied in different domains, going from clinical trials where the goal is investigating the effects of different experimental treatments while minimizing patient losses, to adaptive routing where the goal is to minimize the delays in a network. This article provides a review of the recent results on applying bandit to real-life scenario and summarize the state of the art for each of these fields. Different techniques has been proposed to solve this problem setting, like epsilon-greedy, Upper confident bound (UCB) and Thompson Sampling (TS). We are showing here how this algorithms were adapted to solve the different problems of exploration exploitation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/10/2017

Context Attentive Bandits: Contextual Bandit with Restricted Context

We consider a novel formulation of the multi-armed bandit model, which w...
research
02/25/2014

Algorithms for multi-armed bandit problems

Although many algorithms for the multi-armed bandit problem are well-und...
research
03/04/2019

Learning Modular Safe Policies in the Bandit Setting with Application to Adaptive Clinical Trials

The stochastic multi-armed bandit problem is a well-known model for stud...
research
05/21/2017

Instrument-Armed Bandits

We extend the classic multi-armed bandit (MAB) model to the setting of n...
research
03/17/2019

On Multi-Armed Bandit Designs for Phase I Clinical Trials

We study the problem of finding the optimal dosage in a phase I clinical...
research
10/06/2021

Tuning Confidence Bound for Stochastic Bandits with Bandit Distance

We propose a novel modification of the standard upper confidence bound (...
research
04/21/2023

On the Importance of Exploration for Real Life Learned Algorithms

The quality of data driven learning algorithms scales significantly with...

Please sign up or login with your details

Forgot password? Click here to reset