Asymptotically Optimal Multi-Armed Bandit Activation Policies under Side Constraints

11/30/2018
by   Apostolos N. Burnetas, et al.
0

This paper introduces the first asymptotically optimal strategy for the multi armed bandit (MAB) problem under side constraints. The side constraints model situations in which bandit activations are not cost free, but incur known bandit dependent costs (utilize different resources), and the controller is always constrained by a limited resource availability. The main result involves the derivation of an asymptotic lower bound for the regret of feasible uniformly fast policies and the construction of policies that achieve this lower bound, under pertinent conditions. Further, we provide the explicit form of such policies for the case in which the unknown distributions are Normal with unknown means and known variances and for the case of arbitrary discrete distributions with finite support.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/09/2015

Asymptotically Optimal Multi-Armed Bandit Policies under a Cost Constraint

We develop asymptotically optimal policies for the multi armed bandit (M...
research
05/18/2012

Thompson Sampling: An Asymptotically Optimal Finite Time Analysis

The question of the optimality of Thompson Sampling for solving the stoc...
research
02/03/2023

An Asymptotically Optimal Algorithm for the One-Dimensional Convex Hull Feasibility Problem

This work studies the pure-exploration setting for the convex hull feasi...
research
10/02/2020

On Statistical Discrimination as a Failure of Social Learning: A Multi-Armed Bandit Approach

We analyze statistical discrimination using a multi-armed bandit model w...
research
01/06/2016

On Bayesian index policies for sequential resource allocation

This paper is about index policies for minimizing (frequentist) regret i...
research
04/26/2017

Reward Maximization Under Uncertainty: Leveraging Side-Observations on Networks

We study the stochastic multi-armed bandit (MAB) problem in the presence...
research
02/14/2012

Graphical Models for Bandit Problems

We introduce a rich class of graphical models for multi-armed bandit pro...

Please sign up or login with your details

Forgot password? Click here to reset