A Unified View of Large-scale Zero-sum Equilibrium Computation

11/18/2014
by   Kevin Waugh, et al.
0

The task of computing approximate Nash equilibria in large zero-sum extensive-form games has received a tremendous amount of attention due mainly to the Annual Computer Poker Competition. Immediately after its inception, two competing and seemingly different approaches emerged---one an application of no-regret online learning, the other a sophisticated gradient method applied to a convex-concave saddle-point formulation. Since then, both approaches have grown in relative isolation with advancements on one side not effecting the other. In this paper, we rectify this by dissecting and, in a sense, unify the two views.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/29/2022

No-Regret Learning in Network Stochastic Zero-Sum Games

No-regret learning has been widely used to compute a Nash equilibrium in...
research
02/03/2020

Local Nash Equilibria are Isolated, Strict Local Nash Equilibria in `Almost All' Zero-Sum Continuous Games

We prove that differential Nash equilibria are generic amongst local Nas...
research
07/18/2022

Fast Convergence of Optimistic Gradient Ascent in Network Zero-Sum Extensive Form Games

The study of learning in games has thus far focused primarily on normal ...
research
11/05/2021

Online Learning in Periodic Zero-Sum Games

A seminal result in game theory is von Neumann's minmax theorem, which s...
research
04/01/2022

Robust remote estimation over the collision channel in the presence of an intelligent jammer

We consider a sensor-receiver pair communicating over a wireless channel...
research
07/27/2018

Acceleration through Optimistic No-Regret Dynamics

We consider the problem of minimizing a smooth convex function by reduci...
research
09/08/2017

Cycles in adversarial regularized learning

Regularized learning is a fundamental technique in online optimization, ...

Please sign up or login with your details

Forgot password? Click here to reset