On Finding the Largest Mean Among Many

06/17/2013
by   Kevin Jamieson, et al.
0

Sampling from distributions to find the one with the largest mean arises in a broad range of applications, and it can be mathematically modeled as a multi-armed bandit problem in which each distribution is associated with an arm. This paper studies the sample complexity of identifying the best arm (largest mean) in a multi-armed bandit problem. Motivated by large-scale applications, we are especially interested in identifying situations where the total number of samples that are necessary and sufficient to find the best arm scale linearly with the number of arms. We present a single-parameter multi-armed bandit model that spans the range from linear to superlinear sample complexity. We also give a new algorithm for best arm identification, called PRISM, with linear sample complexity for a wide range of mean distributions. The algorithm, like most exploration procedures for multi-armed bandits, is adaptive in the sense that the next arms to sample are selected based on previous samples. We compare the sample complexity of adaptive procedures with simpler non-adaptive procedures using new lower bounds. For many problem instances, the increased sample complexity required by non-adaptive procedures is a polynomial factor of the number of arms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/14/2018

Sample complexity of partition identification using multi-armed bandits

Given a vector of probability distributions, or arms, each of which can ...
research
02/26/2018

Best Arm Identification for Contaminated Bandits

We propose the Contaminated Best Arm Identification variant of the Multi...
research
12/27/2013

lil' UCB : An Optimal Exploration Algorithm for Multi-Armed Bandits

The paper proposes a novel upper confidence bound (UCB) procedure for id...
research
01/24/2019

PAC Identification of Many Good Arms in Stochastic Multi-Armed Bandits

We consider the problem of identifying any k out of the best m arms in a...
research
06/15/2019

The True Sample Complexity of Identifying Good Arms

We consider two multi-armed bandit problems with n arms: (i) given an ϵ ...
research
01/31/2015

Sparse Dueling Bandits

The dueling bandit problem is a variation of the classical multi-armed b...
research
01/31/2017

Learning the distribution with largest mean: two bandit frameworks

Over the past few years, the multi-armed bandit model has become increas...

Please sign up or login with your details

Forgot password? Click here to reset