On the Heterogeneity of Independent Learning Dynamics in Zero-sum Stochastic Games

12/12/2021
by   Muhammed O. Sayin, et al.
0

We analyze the convergence properties of the two-timescale fictitious play combining the classical fictitious play with the Q-learning for two-player zero-sum stochastic games with player-dependent learning rates. We show its almost sure convergence under the standard assumptions in two-timescale stochastic approximation methods when the discount factor is less than the product of the ratios of player-dependent step sizes. To this end, we formulate a novel Lyapunov function formulation and present a one-sided asynchronous convergence result.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/04/2022

On the Global Convergence of Stochastic Fictitious Play in Stochastic Games with Turn-based Controllers

This paper presents a learning dynamic with almost sure convergence guar...
research
05/23/2022

Fictitious Play in Markov Games with Single Controller

Certain but important classes of strategic-form games, including zero-su...
research
07/10/2020

Exponential Convergence of Gradient Methods in Concave Network Zero-sum Games

Motivated by Generative Adversarial Networks, we study the computation o...
research
11/16/2022

Asynchronous Gradient Play in Zero-Sum Multi-agent Games

Finding equilibria via gradient play in competitive multi-agent games ha...
research
03/19/2023

Instance-dependent Sample Complexity Bounds for Zero-sum Matrix Games

We study the sample complexity of identifying an approximate equilibrium...
research
02/17/2021

Provably Efficient Policy Gradient Methods for Two-Player Zero-Sum Markov Games

Policy gradient methods are widely used in solving two-player zero-sum g...
research
03/03/2023

A Finite-Sample Analysis of Payoff-Based Independent Learning in Zero-Sum Stochastic Games

We study two-player zero-sum stochastic games, and propose a form of ind...

Please sign up or login with your details

Forgot password? Click here to reset