Improved Training of Wasserstein GANs

03/31/2017
by   Ishaan Gulrajani, et al.
0

Generative Adversarial Networks (GANs) are powerful generative models, but suffer from training instability. The recently proposed Wasserstein GAN (WGAN) makes progress toward stable training of GANs, but can still generate low-quality samples or fail to converge in some settings. We find that these problems are often due to the use of weight clipping in WGAN to enforce a Lipschitz constraint on the critic, which can lead to pathological behavior. We propose an alternative to clipping weights: penalize the norm of gradient of the critic with respect to its input. Our proposed method performs better than standard WGAN and enables stable training of a wide variety of GAN architectures with almost no hyperparameter tuning, including 101-layer ResNets and language models over discrete data. We also achieve high quality generations on CIFAR-10 and LSUN bedrooms.

READ FULL TEXT

page 5

page 7

page 8

page 15

page 16

page 17

page 18

page 19

research
07/12/2019

Virtual Adversarial Lipschitz Regularization

Generative adversarial networks (GANs) are one of the most popular appro...
research
03/16/2018

Varying k-Lipschitz Constraint for Generative Adversarial Networks

Generative Adversarial Networks (GANs) are powerful generative models, b...
research
05/06/2017

Face Super-Resolution Through Wasserstein GANs

Generative adversarial networks (GANs) have received a tremendous amount...
research
05/31/2021

The use of Generative Adversarial Networks to characterise new physics in multi-lepton final states at the LHC

Semi-supervision in Machine Learning can be used in searches for new phy...
research
09/19/2017

Summable Reparameterizations of Wasserstein Critics in the One-Dimensional Setting

Generative adversarial networks (GANs) are an exciting alternative to al...
research
12/03/2018

A Wasserstein GAN model with the total variational regularization

It is well known that the generative adversarial nets (GANs) are remarka...
research
10/27/2021

Training Wasserstein GANs without gradient penalties

We propose a stable method to train Wasserstein generative adversarial n...

Please sign up or login with your details

Forgot password? Click here to reset