ODE Analysis of Stochastic Gradient Methods with Optimism and Anchoring for Minimax Problems and GANs

05/26/2019
by   Ernest K. Ryu, et al.
10

Despite remarkable empirical success, the training dynamics of generative adversarial networks (GAN), which involves solving a minimax game using stochastic gradients, is still poorly understood. In this work, we analyze last-iterate convergence of simultaneous gradient descent (simGD) and its variants under the assumption of convex-concavity, guided by a continuous-time analysis with differential equations. First, we show that simGD, as is, converges with stochastic sub-gradients under strict convexity in the primal variable. Second, we generalize optimistic simGD to accommodate an optimism rate separate from the learning rate and show its convergence with full gradients. Finally, we present anchored simGD, a new method, and show convergence with stochastic subgradients.

READ FULL TEXT
research
05/14/2023

Local Convergence of Gradient Descent-Ascent for Training Generative Adversarial Networks

Generative Adversarial Networks (GANs) are a popular formulation to trai...
research
02/25/2021

Local Stochastic Gradient Descent Ascent: Convergence Analysis and Communication Efficiency

Local SGD is a promising approach to overcome the communication overhead...
research
12/18/2020

Convergence dynamics of Generative Adversarial Networks: the dual metric flows

Fitting neural networks often resorts to stochastic (or similar) gradien...
research
10/28/2020

Training Generative Adversarial Networks by Solving Ordinary Differential Equations

The instability of Generative Adversarial Network (GAN) training has fre...
research
06/18/2021

Local AdaGrad-Type Algorithm for Stochastic Convex-Concave Minimax Problems

Large scale convex-concave minimax problems arise in numerous applicatio...
research
12/01/2021

Convergence of GANs Training: A Game and Stochastic Control Methodology

Training of generative adversarial networks (GANs) is known for its diff...
research
11/18/2021

Second-Order Mirror Descent: Convergence in Games Beyond Averaging and Discounting

In this paper, we propose a second-order extension of the continuous-tim...

Please sign up or login with your details

Forgot password? Click here to reset