Tight (Lower) Bounds for the Fixed Budget Best Arm Identification Bandit Problem

05/29/2016
by   Alexandra Carpentier, et al.
0

We consider the problem of best arm identification with a fixed budget T, in the K-armed stochastic bandit setting, with arms distribution defined on [0,1]. We prove that any bandit strategy, for at least one bandit problem characterized by a complexity H, will misidentify the best arm with probability lower bounded by (-T/(K)H), where H is the sum for all sub-optimal arms of the inverse of the squared gaps. Our result disproves formally the general belief - coming from results in the fixed confidence setting - that there must exist an algorithm for this problem whose probability of error is upper bounded by (-T/H). This also proves that some existing strategies based on the Successive Rejection of the arms are optimal - closing therefore the current gap between upper and lower bounds for the fixed budget best arm identification problem.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/16/2023

On the Existence of a Complexity in Fixed Budget Bandit Identification

In fixed budget bandit identification, an algorithm sequentially observe...
research
09/30/2022

On Best-Arm Identification with a Fixed Budget in Non-Parametric Multi-Armed Bandits

We lay the foundations of a non-parametric theory of best-arm identifica...
research
11/15/2022

Bayesian Fixed-Budget Best-Arm Identification

Fixed-budget best-arm identification (BAI) is a bandit problem where the...
research
06/13/2023

Fixed-Budget Best-Arm Identification with Heterogeneous Reward Variances

We study the problem of best-arm identification (BAI) in the fixed-budge...
research
03/02/2023

Open Problem: Optimal Best Arm Identification with Fixed Budget

Best arm identification or pure exploration problems have received much ...
research
10/22/2020

Quantile Bandits for Best Arms Identification with Concentration Inequalities

We consider a variant of the best arm identification task in stochastic ...
research
05/27/2016

An optimal algorithm for the Thresholding Bandit Problem

We study a specific combinatorial pure exploration stochastic bandit pro...

Please sign up or login with your details

Forgot password? Click here to reset