Optimal Fair Multi-Agent Bandits

06/07/2023
by   Amir Leshem, et al.
0

In this paper, we study the problem of fair multi-agent multi-arm bandit learning when agents do not communicate with each other, except collision information, provided to agents accessing the same arm simultaneously. We provide an algorithm with regret O(N^3 log N log T ) (assuming bounded rewards, with unknown bound). This significantly improves previous results which had regret of order O(log T loglog T) and exponential dependence on the number of agents. The result is attained by using a distributed auction algorithm to learn the sample-optimal matching, a new type of exploitation phase whose length is derived from the observed samples, and a novel order-statistics-based regret analysis. Simulation results present the dependence of the regret on log T.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/10/2021

Multi-Agent Multi-Armed Bandits with Limited Communication

We consider the problem where N agents collaboratively interact with an ...
research
12/20/2021

Balancing Adaptability and Non-exploitability in Repeated Games

We study the problem of guaranteeing low regret in repeated games agains...
research
02/28/2022

Robust Multi-Agent Bandits Over Undirected Graphs

We consider a multi-agent multi-armed bandit setting in which n honest a...
research
09/15/2023

Clustered Multi-Agent Linear Bandits

We address in this paper a particular instance of the multi-agent linear...
research
11/10/2021

Multi-Agent Learning for Iterative Dominance Elimination: Formal Barriers and New Algorithms

Dominated actions are natural (and perhaps the simplest possible) multi-...
research
05/31/2022

Near-Optimal Collaborative Learning in Bandits

This paper introduces a general multi-agent bandit model in which each a...
research
02/15/2021

Distributed Online Learning for Joint Regret with Communication Constraints

In this paper we consider a distributed online learning setting for jo...

Please sign up or login with your details

Forgot password? Click here to reset