Bandits with heavy tail

09/08/2012
by   Sébastien Bubeck, et al.
0

The stochastic multi-armed bandit problem is well understood when the reward distributions are sub-Gaussian. In this paper we examine the bandit problem under the weaker assumption that the distributions have moments of order 1+ϵ, for some ϵ∈ (0,1]. Surprisingly, moments of order 2 (i.e., finite variance) are sufficient to obtain regret bounds of the same order as under sub-Gaussian reward distributions. In order to achieve such regret, we define sampling strategies based on refined estimators of the mean such as the truncated empirical mean, Catoni's M-estimator, and the median-of-means estimator. We also derive matching lower bounds that also show that the best achievable regret deteriorates when ϵ <1.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/20/2020

Minimax Policy for Heavy-tailed Multi-armed Bandits

We study the stochastic Multi-Armed Bandit (MAB) problem under worst cas...
research
09/29/2021

Sequential Estimation under Multiple Resources: a Bandit Point of View

The problem of Sequential Estimation under Multiple Resources (SEMR) is ...
research
02/07/2021

Regret Minimization in Heavy-Tailed Bandits

We revisit the classic regret-minimization problem in the stochastic mul...
research
10/25/2018

Almost Optimal Algorithms for Linear Stochastic Bandits with Heavy-Tailed Payoffs

In linear stochastic bandits, it is commonly assumed that payoffs are wi...
research
10/26/2021

Breaking the Moments Condition Barrier: No-Regret Algorithm for Bandits with Super Heavy-Tailed Payoffs

Despite a large amount of effort in dealing with heavy-tailed error in m...
research
10/12/2019

What You See May Not Be What You Get: UCB Bandit Algorithms Robust to ε-Contamination

Motivated by applications of bandit algorithms in education, we consider...
research
06/12/2019

Bootstrapping Upper Confidence Bound

Upper Confidence Bound (UCB) method is arguably the most celebrated one ...

Please sign up or login with your details

Forgot password? Click here to reset