Convergence dynamics of Generative Adversarial Networks: the dual metric flows

12/18/2020
by   Gabriel Turinici, et al.
0

Fitting neural networks often resorts to stochastic (or similar) gradient descent which is a noise-tolerant (and efficient) resolution of a gradient descent dynamics. It outputs a sequence of networks parameters, which sequence evolves during the training steps. The gradient descent is the limit, when the learning rate is small and the batch size is infinite, of this set of increasingly optimal network parameters obtained during training. In this contribution, we investigate instead the convergence in the Generative Adversarial Networks used in machine learning. We study the limit of small learning rate, and show that, similar to single network training, the GAN learning dynamics tend, for vanishing learning rate to some limit dynamics. This leads us to consider evolution equations in metric spaces (which is the natural framework for evolving probability laws) that we call dual flows. We give formal definitions of solutions and prove the convergence. The theory is then applied to specific instances of GANs and we discuss how this insight helps understand and mitigate the mode collapse. Keywords: GAN; metric flow; generative network

READ FULL TEXT

page 1

page 2

page 3

page 4

research
02/14/2023

Stochastic Modified Flows, Mean-Field Limits and Dynamics of Stochastic Gradient Descent

We propose new limiting dynamics for stochastic gradient descent in the ...
research
05/26/2019

ODE Analysis of Stochastic Gradient Methods with Optimism and Anchoring for Minimax Problems and GANs

Despite remarkable empirical success, the training dynamics of generativ...
research
01/19/2020

Dual Stochastic Natural Gradient Descent

Although theoretically appealing, Stochastic Natural Gradient Descent (S...
research
02/16/2018

Interaction Matters: A Note on Non-asymptotic Local Convergence of Generative Adversarial Networks

Motivated by the pursuit of a systematic computational and algorithmic u...
research
10/28/2020

Training Generative Adversarial Networks by Solving Ordinary Differential Equations

The instability of Generative Adversarial Network (GAN) training has fre...
research
02/22/2022

A duality connecting neural network and cosmological dynamics

We demonstrate that the dynamics of neural networks trained with gradien...
research
10/05/2022

Non-Convergence and Limit Cycles in the Adam optimizer

One of the most popular training algorithms for deep neural networks is ...

Please sign up or login with your details

Forgot password? Click here to reset