Generalization in Mean Field Games by Learning Master Policies

09/20/2021
by   Sarah Perrin, et al.
0

Mean Field Games (MFGs) can potentially scale multi-agent systems to extremely large populations of agents. Yet, most of the literature assumes a single initial distribution for the agents, which limits the practical applications of MFGs. Machine Learning has the potential to solve a wider diversity of MFG problems thanks to generalizations capacities. We study how to leverage these generalization properties to learn policies enabling a typical agent to behave optimally against any population distribution. In reference to the Master equation in MFGs, we coin the term “Master policies” to describe them and we prove that a single Master policy provides a Nash equilibrium, whatever the initial distribution. We propose a method to learn such Master policies. Our approach relies on three ingredients: adding the current population distribution as part of the observation, approximating Master policies with neural networks, and training via Reinforcement Learning and Fictitious Play. We illustrate on numerical examples not only the efficiency of the learned Master policy but also its generalization capabilities beyond the distributions used for training.

READ FULL TEXT

page 9

page 14

research
12/29/2022

Policy Mirror Ascent for Efficient and Independent Learning in Mean Field Games

Mean-field games have been used as a theoretical tool to obtain an appro...
research
04/04/2023

Regularization of the policy updates for stabilizing Mean Field Games

This work studies non-cooperative Multi-Agent Reinforcement Learning (MA...
research
06/26/2023

On Imitation in Mean-field Games

We explore the problem of imitation learning (IL) in the context of mean...
research
10/09/2018

Collective Strategies with a Master-slave Mechanism Dominate in Spatial Iterated Prisoner's Dilemma

The Iterated Prisoner's Dilemma (IPD) can model transactions among indiv...
research
02/28/2021

Scaling up Mean Field Games with Online Mirror Descent

We address scaling up equilibrium computation in Mean Field Games (MFGs)...
research
05/31/2022

Simplex NeuPL: Any-Mixture Bayes-Optimality in Symmetric Zero-sum Games

Learning to play optimally against any mixture over a diverse set of str...
research
09/13/2022

Independent Learning in Mean-Field Games: Satisficing Paths and Convergence to Subjective Equilibria

Independent learners are learning agents that naively employ single-agen...

Please sign up or login with your details

Forgot password? Click here to reset