Policy Mirror Ascent for Efficient and Independent Learning in Mean Field Games

12/29/2022
∙
by   Batuhan Yardim, et al.
∙
0
∙

Mean-field games have been used as a theoretical tool to obtain an approximate Nash equilibrium for symmetric and anonymous N-player games in literature. However, limiting applicability, existing theoretical results assume variations of a "population generative model", which allows arbitrary modifications of the population distribution by the learning algorithm. Instead, we show that N agents running policy mirror ascent converge to the Nash equilibrium of the regularized game within 𝒊Ėƒ(Îĩ^-2) samples from a single sample trajectory without a population generative model, up to a standard 𝒊(1/√(N)) error due to the mean field. Taking a divergent approach from literature, instead of working with the best-response map we first show that a policy mirror ascent map can be used to construct a contractive operator having the Nash equilibrium as its fixed point. Next, we prove that conditional TD-learning in N-agent games can learn value functions within 𝒊Ėƒ(Îĩ^-2) time steps. These results allow proving sample complexity guarantees in the oracle-free setting by only relying on a sample path from the N agent simulator. Furthermore, we demonstrate that our methodology allows for independent learning by N agents with finite sample guarantees.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
∙ 10/14/2021

Shaping Large Population Agent Behaviors Through Entropy-Regularized Mean-Field Games

Mean-field games (MFG) were introduced to efficiently analyze approximat...
research
∙ 09/20/2021

Generalization in Mean Field Games by Learning Master Policies

Mean Field Games (MFGs) can potentially scale multi-agent systems to ext...
research
∙ 11/25/2020

Exploratory LQG Mean Field Games with Entropy Regularization

We study a general class of entropy-regularized multi-variate LQG mean f...
research
∙ 06/05/2023

Networked Communication for Decentralised Agents in Mean-Field Games

We introduce networked communication to the mean-field game framework. I...
research
∙ 05/18/2023

On the Statistical Efficiency of Mean Field Reinforcement Learning with General Function Approximation

In this paper, we study the statistical efficiency of Reinforcement Lear...
research
∙ 02/28/2021

Scaling up Mean Field Games with Online Mirror Descent

We address scaling up equilibrium computation in Mean Field Games (MFGs)...
research
∙ 05/19/2023

Monte-Carlo Search for an Equilibrium in Dec-POMDPs

Decentralized partially observable Markov decision processes (Dec-POMDPs...

Please sign up or login with your details

Forgot password? Click here to reset