Finding All ε-Good Arms in Stochastic Bandits

06/16/2020
by   Blake Mason, et al.
0

The pure-exploration problem in stochastic multi-armed bandits aims to find one or more arms with the largest (or near largest) means. Examples include finding an ϵ-good arm, best-arm identification, top-k arm identification, and finding all arms with means above a specified threshold. However, the problem of finding all ϵ-good arms has been overlooked in past work, although arguably this may be the most natural objective in many applications. For example, a virologist may conduct preliminary laboratory experiments on a large candidate set of treatments and move all ϵ-good treatments into more expensive clinical trials. Since the ultimate clinical efficacy is uncertain, it is important to identify all ϵ-good candidates. Mathematically, the all-ϵ-good arm identification problem presents significant new challenges and surprises that do not arise in the pure-exploration objectives studied in the past. We introduce two algorithms to overcome these and demonstrate their great empirical performance on a large-scale crowd-sourced dataset of 2.2M ratings collected by the New Yorker Caption Contest as well as a dataset testing hundreds of possible cancer drugs.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/19/2017

Practical Algorithms for Best-K Identification in Multi-Armed Bandits

In the Best-K identification problem (Best-K-Arm), we are given N stocha...
research
03/13/2023

Differential Good Arm Identification

This paper targets a variant of the stochastic multi-armed bandit proble...
research
12/08/2021

Best Arm Identification under Additive Transfer Bandits

We consider a variant of the best arm identification (BAI) problem in mu...
research
08/02/2021

Pure Exploration in Multi-armed Bandits with Graph Side Information

We study pure exploration in multi-armed bandits with graph side-informa...
research
06/13/2022

Top Two Algorithms Revisited

Top Two algorithms arose as an adaptation of Thompson sampling to best a...
research
11/15/2018

Pure-Exploration for Infinite-Armed Bandits with General Arm Reservoirs

This paper considers a multi-armed bandit game where the number of arms ...
research
06/03/2019

MaxGap Bandit: Adaptive Algorithms for Approximate Ranking

This paper studies the problem of adaptively sampling from K distributio...

Please sign up or login with your details

Forgot password? Click here to reset