Double Descent and Other Interpolation Phenomena in GANs

06/07/2021
by   Lorenzo Luzi, et al.
0

We study overparameterization in generative adversarial networks (GANs) that can interpolate the training data. We show that overparameterization can improve generalization performance and accelerate the training process. We study the generalization error as a function of latent space dimension and identify two main behaviors, depending on the learning setting. First, we show that overparameterized generative models that learn distributions by minimizing a metric or f-divergence do not exhibit double descent in generalization errors; specifically, all the interpolating solutions achieve the same generalization error. Second, we develop a new pseudo-supervised learning approach for GANs where the training utilizes pairs of fabricated (noise) inputs in conjunction with real output samples. Our pseudo-supervised setting exhibits double descent (and in some cases, triple descent) of generalization errors. We combine pseudo-supervision with overparameterization (i.e., overly large latent space dimension) to accelerate training while performing better, or close to, the generalization performance without pseudo-supervision. While our analysis focuses mostly on linear GANs, we also apply important insights for improving generalization of nonlinear, multilayer GANs.

READ FULL TEXT

page 10

page 24

research
11/29/2020

Generalization and Memorization: The Bias Potential Model

Models for learning probability distributions such as generative models ...
research
02/25/2020

Subspace Fitting Meets Regression: The Effects of Supervision and Orthonormality Constraints on Double Descent of Generalization Errors

We study the linear subspace fitting problem in the overparameterized se...
research
07/21/2022

Optimal precision for GANs

When learning disconnected distributions, Generative adversarial network...
research
08/26/2021

When and how epochwise double descent happens

Deep neural networks are known to exhibit a `double descent' behavior as...
research
07/06/2023

A Privacy-Preserving Walk in the Latent Space of Generative Models for Medical Applications

Generative Adversarial Networks (GANs) have demonstrated their ability t...
research
10/30/2017

Understanding GANs: the LQG Setting

Generative Adversarial Networks (GANs) have become a popular method to l...
research
03/10/2023

Unifying Grokking and Double Descent

A principled understanding of generalization in deep learning may requir...

Please sign up or login with your details

Forgot password? Click here to reset