O(1/T) Time-Average Convergence in a Generalization of Multiagent Zero-Sum Games

10/06/2021
by   James P. Bailey, et al.
0

We introduce a generalization of zero-sum network multiagent matrix games and prove that alternating gradient descent converges to the set of Nash equilibria at rate O(1/T) for this set of games. Alternating gradient descent obtains this convergence guarantee while using fixed learning rates that are four times larger than the optimistic variant of gradient descent. Experimentally, we show with 97.5 time-averaged strategies that are 2.585 times closer to the set of Nash equilibria than optimistic gradient descent.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/18/2022

Fast Convergence of Optimistic Gradient Ascent in Network Zero-Sum Extensive Form Games

The study of learning in games has thus far focused primarily on normal ...
research
02/08/2021

Last-iterate Convergence of Decentralized Optimistic Gradient Descent/Ascent in Infinite-horizon Competitive Markov Games

We study infinite-horizon discounted two-player zero-sum Markov games, a...
research
09/30/2020

Solving Zero-Sum Games through Alternating Projections

In this work, we establish near-linear and strong convergence for a natu...
research
11/16/2021

Polymatrix Competitive Gradient Descent

Many economic games and machine learning approaches can be cast as compe...
research
05/27/2022

Competitive Gradient Optimization

We study the problem of convergence to a stationary point in zero-sum ga...
research
05/28/2019

Competitive Gradient Descent

We introduce a new algorithm for the numerical computation of Nash equil...
research
06/08/2022

Alternating Mirror Descent for Constrained Min-Max Games

In this paper we study two-player bilinear zero-sum games with constrain...

Please sign up or login with your details

Forgot password? Click here to reset