Extra-gradient with player sampling for provable fast convergence in n-player games

05/29/2019
by   Samy Jelassi, et al.
0

Data-driven model training is increasingly relying on finding Nash equilibria with provable techniques, e.g., for GANs and multi-agent RL. In this paper, we analyse a new extra-gradient method, that performs gradient extrapolations and updates on a random subset of players at each iteration. This approach provably exhibits the same rate of convergence as full extra-gradient in non-smooth convex games. We propose an additional variance reduction mechanism for this to hold for smooth convex games. Our approach makes extrapolation amenable to massive multiplayer settings, and brings empirical speed-ups, in particular when using cyclic sampling schemes. We demonstrate the efficiency of player sampling on large-scale non-smooth and non-strictly convex games. We show that the joint use of extrapolation and player sampling allows to train better GANs on CIFAR10.

READ FULL TEXT
research
08/12/2022

Three-Player Game Training Dynamics

This work explores three-player game training dynamics, under what condi...
research
08/22/2022

Minimax-Optimal Multi-Agent RL in Markov Games With a Generative Model

This paper studies multi-agent reinforcement learning in Markov games, w...
research
01/14/2020

Smooth markets: A basic mechanism for organizing gradient-based learners

With the success of modern machine learning, it is becoming increasingly...
research
11/10/2021

The Cut and Play Algorithm: Computing Nash Equilibria via Outer Approximations

The concept of Nash equilibrium enlightens the structure of rational beh...
research
10/26/2020

Tight last-iterate convergence rates for no-regret learning in multi-player games

We study the question of obtaining last-iterate convergence rates for no...
research
06/08/2020

Hedging in games: Faster convergence of external and swap regrets

We consider the setting where players run the Hedge algorithm or its opt...
research
08/13/2018

Fast, Better Training Trick -- Random Gradient

In this paper, we will show an unprecedented method to accelerate traini...

Please sign up or login with your details

Forgot password? Click here to reset