Multi-Player Bandits: The Adversarial Case

02/21/2019
by   Pragnya Alatur, et al.
0

We consider a setting where multiple players sequentially choose among a common set of actions (arms). Motivated by a cognitive radio networks application, we assume that players incur a loss upon colliding, and that communication between players is not possible. Existing approaches assume that the system is stationary. Yet this assumption is often violated in practice, e.g., due to signal strength fluctuations. In this work, we design the first Multi-player Bandit algorithm that provably works in arbitrarily changing environments, where the losses of the arms may even be chosen by an adversary. This resolves an open problem posed by Rosenski, Shamir, and Szlak (2016).

READ FULL TEXT
research
10/26/2018

Game of Thrones: Fully Distributed Learning for Multi-Player Bandits

We consider a multi-armed bandit game where N players compete for M arms...
research
10/26/2018

Distributed Multi-Player Bandits - a Game of Thrones Approach

We consider a multi-armed bandit game where N players compete for K arms...
research
12/09/2015

Multi-Player Bandits -- a Musical Chairs Approach

We consider a variant of the stochastic multi-armed bandit problem, wher...
research
05/30/2023

Competing for Shareable Arms in Multi-Player Multi-Armed Bandits

Competitions for shareable and limited resources have long been studied ...
research
11/29/2022

A survey on multi-player bandits

Due mostly to its application to cognitive radio networks, multiplayer b...
research
02/23/2020

My Fair Bandit: Distributed Learning of Max-Min Fairness with Multi-player Bandits

Consider N cooperative but non-communicating players where each plays on...
research
09/18/2023

How to Make Knockout Tournaments More Popular?

Given a mapping from a set of players to the leaves of a complete binary...

Please sign up or login with your details

Forgot password? Click here to reset