On Distributed Cooperative Decision-Making in Multiarmed Bandits

12/21/2015
by   Peter Landgren, et al.
0

We study the explore-exploit tradeoff in distributed cooperative decision-making using the context of the multiarmed bandit (MAB) problem. For the distributed cooperative MAB problem, we design the cooperative UCB algorithm that comprises two interleaved distributed processes: (i) running consensus algorithms for estimation of rewards, and (ii) upper-confidence-bound-based heuristics for selection of arms. We rigorously analyze the performance of the cooperative UCB algorithm and characterize the influence of communication graph structure on the decision-making performance of the group.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/03/2020

Distributed Cooperative Decision Making in Multi-agent Multi-armed Bandits

We study a distributed decision-making problem in which multiple agents ...
research
10/08/2021

When to Call Your Neighbor? Strategic Communication in Cooperative Stochastic Bandits

In cooperative bandits, a framework that captures essential features of ...
research
07/05/2015

Correlated Multiarmed Bandit Problem: Bayesian Algorithms and Regret Analysis

We consider the correlated multiarmed bandit (MAB) problem in which the ...
research
04/13/2020

Distributed Learning: Sequential Decision Making in Resource-Constrained Environments

We study cost-effective communication strategies that can be used to imp...
research
12/14/2018

Individual decision making in task-oriented groups

The strategies adopted by individuals to select relevant information to ...
research
05/02/2019

Towards a Novel Cooperative Logistics Information System Framework

Supply Chains and Logistics have a growing importance in global economy....
research
08/03/2018

A Cooperative Group Optimization System

A cooperative group optimization (CGO) system is presented to implement ...

Please sign up or login with your details

Forgot password? Click here to reset