The Power of Regularization in Solving Extensive-Form Games

06/19/2022
by   Mingyang Liu, et al.
0

In this paper, we investigate the power of regularization, a common technique in reinforcement learning and optimization, in solving extensive-form games (EFGs). We propose a series of new algorithms based on regularizing the payoff functions of the game, and establish a set of convergence results that strictly improve over the existing ones, with either weaker assumptions or stronger convergence guarantees. In particular, we first show that dilated optimistic mirror descent (DOMD), an efficient variant of OMD for solving EFGs, with adaptive regularization can achieve a fast Õ(1/T) last-iterate convergence in terms of duality gap without the uniqueness assumption of the Nash equilibrium (NE). Moreover, regularized dilated optimistic multiplicative weights update (Reg-DOMWU), an instance of Reg-DOMD, further enjoys the Õ(1/T) last-iterate convergence rate of the distance to the set of NE. This addresses an open question on whether iterate convergence can be obtained for OMWU algorithms without the uniqueness assumption in both the EFG and normal-form game literature. Second, we show that regularized counterfactual regret minimization (Reg-CFR), with a variant of optimistic mirror descent algorithm as regret-minimizer, can achieve O(1/T^1/4) best-iterate, and O(1/T^3/4) average-iterate convergence rate for finding NE in EFGs. Finally, we show that Reg-CFR can achieve asymptotic last-iterate convergence, and optimal O(1/T) average-iterate convergence rate, for finding the NE of perturbed EFGs, which is useful for finding approximate extensive-form perfect equilibria (EFPE). To the best of our knowledge, they constitute the first last-iterate convergence results for CFR-type algorithms, while matching the SOTA average-iterate convergence rate in finding NE for non-perturbed EFGs. We also provide numerical results to corroborate the advantages of our algorithms.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/09/2017

Regret Minimization in Behaviorally-Constrained Zero-Sum Games

No-regret learning has emerged as a powerful tool for solving extensive-...
research
10/24/2019

Optimistic Regret Minimization for Extensive-Form Games via Dilated Distance-Generating Functions

We study the performance of optimistic regret-minimization algorithms fo...
research
04/11/2022

Equilibrium Finding in Normal-Form Games Via Greedy Regret Minimization

We extend the classic regret minimization framework for approximating eq...
research
06/06/2022

Policy Optimization for Markov Games: Unified Framework and Faster Convergence

This paper studies policy optimization algorithms for multi-agent reinfo...
research
06/07/2021

Forward Looking Best-Response Multiplicative Weights Update Methods

We propose a novel variant of the multiplicative weights update method w...
research
02/13/2019

Stable-Predictive Optimistic Counterfactual Regret Minimization

The CFR framework has been a powerful tool for solving large-scale exten...
research
12/20/2022

Anticipatory Fictitious Play

Fictitious play is an algorithm for computing Nash equilibria of matrix ...

Please sign up or login with your details

Forgot password? Click here to reset