Thresholding Bandit for Dose-ranging: The Impact of Monotonicity

11/13/2017
by   Aurélien Garivier, et al.
0

We analyze the sample complexity of the thresholding bandit problem, with and without the assumption that the mean values of the arms are increasing. In each case, we provide a lower bound valid for any risk δ and any δ-correct algorithm; in addition, we propose an algorithm whose sample complexity is of the same order of magnitude for small risks. This work is motivated by phase 1 clinical trials, a practically important setting where the arm means are increasing by nature, and where no satisfactory solution is available so far.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/13/2017

Nearly Instance Optimal Sample Complexity Bounds for Top-k Arm Selection

In the Best-k-Arm problem, we are given n stochastic bandit arms, each a...
research
06/20/2020

An Optimal Elimination Algorithm for Learning a Best Arm

We consider the classic problem of (ϵ,δ)-PAC learning a best arm where t...
research
05/20/2020

Best Arm Identification in Spectral Bandits

We study best-arm identification with fixed confidence in bandit models ...
research
08/22/2016

Towards Instance Optimal Bounds for Best Arm Identification

In the classical best arm identification (Best-1-Arm) problem, we are gi...
research
12/23/2015

The Max K-Armed Bandit: PAC Lower Bounds and Efficient Algorithms

We consider the Max K-Armed Bandit problem, where a learning agent is fa...
research
02/09/2019

Pure Exploration with Multiple Correct Answers

We determine the sample complexity of pure exploration bandit problems w...
research
10/14/2019

Thresholding Bandit Problem with Both Duels and Pulls

The Thresholding Bandit Problem (TBP) aims to find the set of arms with ...

Please sign up or login with your details

Forgot password? Click here to reset