A Dynamic Observation Strategy for Multi-agent Multi-armed Bandit Problem

04/08/2020
by   Udari Madhushani, et al.
0

We define and analyze a multi-agent multi-armed bandit problem in which decision-making agents can observe the choices and rewards of their neighbors under a linear observation cost. Neighbors are defined by a network graph that encodes the inherent observation constraints of the system. We define a cost associated with observations such that at every instance an agent makes an observation it receives a constant observation regret. We design a sampling algorithm and an observation protocol for each agent to maximize its own expected cumulative reward through minimizing expected cumulative sampling regret and expected cumulative observation regret. For our proposed protocol, we prove that total cumulative regret is logarithmically bounded. We verify the accuracy of analytical bounds using numerical simulations.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/21/2019

Heterogeneous Stochastic Interactions for Multiple Agents in a Multi-armed Bandit Problem

We define and analyze a multi-agent multi-armed bandit problem in which ...
research
10/20/2020

Bayesian Algorithms for Decentralized Stochastic Bandits

We study a decentralized cooperative multi-agent multi-armed bandit prob...
research
10/07/2019

An Option and Agent Selection Policy with Logarithmic Regret for Multi Agent Multi Armed Bandit Problems on Random Graphs

Existing studies of the Multi Agent Multi Armed Bandit (MAMAB) problem, ...
research
06/07/2022

The Survival Bandit Problem

We study the survival bandit problem, a variant of the multi-armed bandi...
research
09/15/2018

Incorporating Behavioral Constraints in Online AI Systems

AI systems that learn through reward feedback about the actions they tak...
research
08/13/2023

Estimating and Incentivizing Imperfect-Knowledge Agents with Hidden Rewards

In practice, incentive providers (i.e., principals) often cannot observe...

Please sign up or login with your details

Forgot password? Click here to reset