Training GANs with Optimism

10/31/2017
by   Constantinos Daskalakis, et al.
0

We address the issue of limit cycling behavior in training Generative Adversarial Networks and propose the use of Optimistic Mirror Decent (OMD) for training Wasserstein GANs. Recent theoretical results have shown that optimistic mirror decent (OMD) can enjoy faster regret rates in the context of zero-sum games. WGANs is exactly a context of solving a zero-sum game with simultaneous no-regret dynamics. Moreover, we show that optimistic mirror decent addresses the limit cycling problem in training WGANs. We formally show that in the case of bi-linear zero-sum games the last iterate of OMD dynamics converges to an equilibrium, in contrast to GD dynamics which are bound to cycle. We also portray the huge qualitative difference between GD and OMD dynamics with toy examples, even when GD is modified with many adaptations proposed in the recent literature, such as gradient penalty or momentum. We apply OMD WGAN training to a bioinformatics problem of generating DNA sequences. We observe that models trained with OMD achieve consistently smaller KL divergence with respect to the true underlying distribution, than models trained with GD variants. Finally, we introduce a new algorithm, Optimistic Adam, which is an optimistic variant of Adam. We apply it to WGAN training on CIFAR10 and observe improved performance in terms of inception score as compared to Adam.

READ FULL TEXT

page 9

page 17

page 27

page 28

page 29

research
03/22/2022

On Last-Iterate Convergence Beyond Zero-Sum Games

Most existing results about last-iterate convergence of learning dynamic...
research
07/27/2018

Acceleration through Optimistic No-Regret Dynamics

We consider the problem of minimizing a smooth convex function by reduci...
research
02/12/2018

Let's be honest: An optimal no-regret framework for zero-sum games

We revisit the problem of solving two-player zero-sum games in the decen...
research
07/12/2018

Negative Momentum for Improved Game Dynamics

Games generalize the optimization paradigm by introducing different obje...
research
02/15/2022

Provably convergent quasistatic dynamics for mean-field two-player zero-sum games

In this paper, we study the problem of finding mixed Nash equilibrium fo...
research
02/16/2021

Complex Momentum for Learning in Games

We generalize gradient descent with momentum for learning in differentia...
research
10/29/2022

Recursive Reasoning in Minimax Games: A Level k Gradient Play Method

Despite the success of generative adversarial networks (GANs) in generat...

Please sign up or login with your details

Forgot password? Click here to reset