Tight last-iterate convergence rates for no-regret learning in multi-player games

10/26/2020
by   Noah Golowich, et al.
0

We study the question of obtaining last-iterate convergence rates for no-regret learning algorithms in multi-player games. We show that the optimistic gradient (OG) algorithm with a constant step-size, which is no-regret, achieves a last-iterate rate of O(1/√(T)) with respect to the gap function in smooth monotone games. This result addresses a question of Mertikopoulos Zhou (2018), who asked whether extra-gradient approaches (such as OG) can be applied to achieve improved guarantees in the multi-agent learning setting. The proof of our upper bound uses a new technique centered around an adaptive choice of potential function at each iteration. We also show that the O(1/√(T)) rate is tight for all p-SCLI algorithms, which includes OG as a special case. As a byproduct of our lower bound analysis we additionally present a proof of a conjecture of Arjevani et al. (2015) which is more direct than previous approaches.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
01/30/2023

Doubly Optimal No-Regret Learning in Monotone Games

We consider online learning in multi-player smooth monotone games. Exist...
research
04/24/2020

Proving μ>1

Choosing the right selection rate is a long standing issue in evolutiona...
research
02/23/2020

Finite-Time Last-Iterate Convergence for Multi-Agent Learning in Games

We consider multi-agent learning via online gradient descent (OGD) in a ...
research
07/29/2023

A new Gradient TD Algorithm with only One Step-size: Convergence Rate Analysis using L-λ Smoothness

Gradient Temporal Difference (GTD) algorithms (Sutton et al., 2008, 2009...
research
05/17/2018

Faster Rates for Convex-Concave Games

We consider the use of no-regret algorithms to compute equilibria for pa...
research
05/29/2019

Extra-gradient with player sampling for provable fast convergence in n-player games

Data-driven model training is increasingly relying on finding Nash equil...
research
06/16/2020

Linear Last-iterate Convergence for Matrix Games and Stochastic Games

Optimistic Gradient Descent Ascent (OGDA) algorithm for saddle-point opt...

Please sign up or login with your details

Forgot password? Click here to reset