Complexity Analysis of a Countable-armed Bandit Problem

01/18/2023
by   Anand Kalvit, et al.
0

We consider a stochastic multi-armed bandit (MAB) problem motivated by “large” action spaces, and endowed with a population of arms containing exactly K arm-types, each characterized by a distinct mean reward. The decision maker is oblivious to the statistical properties of reward distributions as well as the population-level distribution of different arm-types, and is precluded also from observing the type of an arm after play. We study the classical problem of minimizing the expected cumulative regret over a horizon of play n, and propose algorithms that achieve a rate-optimal finite-time instance-dependent regret of 𝒪( log n ). We also show that the instance-independent (minimax) regret is 𝒪̃( √(n)) when K=2. While the order of regret and complexity of the problem suggests a great degree of similarity to the classical MAB problem, properties of the performance bounds and salient aspects of algorithm design are quite distinct from the latter, as are the key primitives that determine complexity along with the analysis tools needed to study them.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/22/2021

From Finite to Countable-Armed Bandits

We consider a stochastic bandit problem with countably many arms that be...
research
06/03/2021

A Closer Look at the Worst-case Behavior of Multi-armed Bandit Algorithms

One of the key drivers of complexity in the classical (stochastic) multi...
research
10/31/2019

Recovering Bandits

We study the recovering bandits problem, a variant of the stochastic mul...
research
06/14/2023

Decentralized Learning Dynamics in the Gossip Model

We study a distributed multi-armed bandit setting among a population of ...
research
04/21/2020

Algorithms for slate bandits with non-separable reward functions

In this paper, we study a slate bandit problem where the function that d...
research
11/15/2022

On Penalization in Stochastic Multi-armed Bandits

We study an important variant of the stochastic multi-armed bandit (MAB)...
research
10/23/2021

The Countable-armed Bandit with Vanishing Arms

We consider a bandit problem with countably many arms, partitioned into ...

Please sign up or login with your details

Forgot password? Click here to reset