Correlated Multiarmed Bandit Problem: Bayesian Algorithms and Regret Analysis

07/05/2015
by   Vaibhav Srivastava, et al.
0

We consider the correlated multiarmed bandit (MAB) problem in which the rewards associated with each arm are modeled by a multivariate Gaussian random variable, and we investigate the influence of the assumptions in the Bayesian prior on the performance of the upper credible limit (UCL) algorithm and a new correlated UCL algorithm. We rigorously characterize the influence of accuracy, confidence, and correlation scale in the prior on the decision-making performance of the algorithms. Our results show how priors and correlation structure can be leveraged to improve performance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/23/2013

Modeling Human Decision-making in Generalized Gaussian Multi-armed Bandits

We present a formal model of human decision-making in explore-exploit ta...
research
12/21/2015

On Distributed Cooperative Decision-Making in Multiarmed Bandits

We study the explore-exploit tradeoff in distributed cooperative decisio...
research
05/30/2018

Infinite Arms Bandit: Optimality via Confidence Bounds

The infinite arms bandit problem was initiated by Berry et al. (1997). T...
research
07/12/2021

Metalearning Linear Bandits by Prior Update

Fully Bayesian approaches to sequential decision-making assume that prob...
research
02/04/2014

Online Stochastic Optimization under Correlated Bandit Feedback

In this paper we consider the problem of online stochastic optimization ...
research
03/06/2023

Thompson Sampling for Linear Bandit Problems with Normal-Gamma Priors

We consider Thompson sampling for linear bandit problems with finitely m...
research
02/08/2021

Correlated Bandits for Dynamic Pricing via the ARC algorithm

The Asymptotic Randomised Control (ARC) algorithm provides a rigorous ap...

Please sign up or login with your details

Forgot password? Click here to reset