Problem Dependent View on Structured Thresholding Bandit Problems

06/18/2021
by   James Cheshire, et al.
0

We investigate the problem dependent regime in the stochastic Thresholding Bandit problem (TBP) under several shape constraints. In the TBP, the objective of the learner is to output, at the end of a sequential game, the set of arms whose means are above a given threshold. The vanilla, unstructured, case is already well studied in the literature. Taking K as the number of arms, we consider the case where (i) the sequence of arm's means (μ_k)_k=1^K is monotonically increasing (MTBP) and (ii) the case where (μ_k)_k=1^K is concave (CTBP). We consider both cases in the problem dependent regime and study the probability of error - i.e. the probability to mis-classify at least one arm. In the fixed budget setting, we provide upper and lower bounds for the probability of error in both the concave and monotone settings, as well as associated algorithms. In both settings the bounds match in the problem dependent regime up to universal constants in the exponential.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/17/2020

The Influence of Shape Constraints on the Thresholding Bandit Problem

We investigate the stochastic Thresholding Bandit problem (TBP) under se...
research
05/27/2016

An optimal algorithm for the Thresholding Bandit Problem

We study a specific combinatorial pure exploration stochastic bandit pro...
research
10/14/2019

Thresholding Bandit Problem with Both Duels and Pulls

The Thresholding Bandit Problem (TBP) aims to find the set of arms with ...
research
02/09/2022

Finding Optimal Arms in Non-stochastic Combinatorial Bandits with Semi-bandit Feedback and Finite Budget

We consider the combinatorial bandits problem with semi-bandit feedback ...
research
11/14/2018

Sample complexity of partition identification using multi-armed bandits

Given a vector of probability distributions, or arms, each of which can ...
research
10/04/2022

Max-Quantile Grouped Infinite-Arm Bandits

In this paper, we consider a bandit problem in which there are a number ...
research
03/09/2016

Best-of-K Bandits

This paper studies the Best-of-K Bandit game: At each time the player ch...

Please sign up or login with your details

Forgot password? Click here to reset