Multiplayer bandits without observing collision information

08/25/2018
by   Gábor Lugosi, et al.
0

We study multiplayer stochastic multi-armed bandit problems in which the players cannot communicate, and if two or more players pull the same arm, a collision occurs and the involved players receive zero reward. We consider two feedback models: a model in which the players can observe whether a collision has occurred, and a more difficult setup when no collision information is available. We give the first theoretical guarantees for the second model: an algorithm with a logarithmic regret, and an algorithm with a square-root regret type that does not depend on the gaps between the means. For the first model, we give the first square-root regret bounds that do not depend on the gaps. Building on these ideas, we also give an algorithm for reaching approximate Nash equilibria quickly in stochastic anti-coordination games.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/04/2019

New Algorithms for Multiplayer Bandits when Arm Means Vary Among Players

We study multiplayer stochastic multi-armed bandit problems in which the...
research
09/28/2019

An Optimal Algorithm in Multiplayer Multi-Armed Bandits

The paper addresses the Multiplayer Multi-Armed Bandit (MMAB) problem, w...
research
04/28/2019

Non-Stochastic Multi-Player Multi-Armed Bandits: Optimal Rate With Collision Information, Sublinear Without

We consider the non-stochastic version of the (cooperative) multi-player...
research
09/21/2018

SIC-MMAB: Synchronisation Involves Communication in Multiplayer Multi-Armed Bandits

We consider the stochastic multiplayer multi-armed bandit problem, where...
research
09/17/2018

Multi-Player Bandits: A Trekking Approach

We study stochastic multi-armed bandits with many players. The players d...
research
02/04/2020

Selfish Robustness and Equilibria in Multi-Player Bandits

Motivated by cognitive radios, stochastic multi-player multi-armed bandi...
research
11/02/2020

On No-Sensing Adversarial Multi-player Multi-armed Bandits with Collision Communications

We study the notoriously difficult no-sensing adversarial multi-player m...

Please sign up or login with your details

Forgot password? Click here to reset