Tackling the Generative Learning Trilemma with Denoising Diffusion GANs

12/15/2021
by   Zhisheng Xiao, et al.
0

A wide variety of deep generative models has been developed in the past decade. Yet, these models often struggle with simultaneously addressing three key requirements including: high sample quality, mode coverage, and fast sampling. We call the challenge imposed by these requirements the generative learning trilemma, as the existing models often trade some of them for others. Particularly, denoising diffusion models have shown impressive sample quality and diversity, but their expensive sampling does not yet allow them to be applied in many real-world applications. In this paper, we argue that slow sampling in these models is fundamentally attributed to the Gaussian assumption in the denoising step which is justified only for small step sizes. To enable denoising with large steps, and hence, to reduce the total number of denoising steps, we propose to model the denoising distribution using a complex multimodal distribution. We introduce denoising diffusion generative adversarial networks (denoising diffusion GANs) that model each denoising step using a multimodal conditional GAN. Through extensive evaluations, we show that denoising diffusion GANs obtain sample quality and diversity competitive with original diffusion models while being 2000× faster on the CIFAR-10 dataset. Compared to traditional GANs, our model exhibits better mode coverage and sample diversity. To the best of our knowledge, denoising diffusion GAN is the first model that reduces sampling cost in diffusion models to an extent that allows them to be applied to real-world applications inexpensively. Project page and code: https://nvlabs.github.io/denoising-diffusion-gan

READ FULL TEXT

page 9

page 22

page 23

page 24

page 25

page 26

page 27

page 28

research
08/28/2023

Voice Conversion with Denoising Diffusion Probabilistic GAN Models

Voice conversion is a method that allows for the transformation of speak...
research
06/29/2022

SPI-GAN: Distilling Score-based Generative Models with Straight-Path Interpolations

Score-based generative models (SGMs) are a recently proposed paradigm fo...
research
05/25/2023

Parallel Sampling of Diffusion Models

Diffusion models are powerful generative models but suffer from slow sam...
research
12/01/2022

Unite and Conquer: Cross Dataset Multimodal Synthesis using Diffusion Models

Generating photos satisfying multiple constraints find broad utility in ...
research
07/05/2023

DiffFlow: A Unified SDE Framework for Score-Based Diffusion Models and Generative Adversarial Networks

Generative models can be categorized into two types: explicit generative...
research
02/12/2019

Learning Generative Models of Structured Signals from Their Superposition Using GANs with Application to Denoising and Demixing

Recently, Generative Adversarial Networks (GANs) have emerged as a popul...
research
03/08/2022

Dynamic Dual-Output Diffusion Models

Iterative denoising-based generation, also known as denoising diffusion ...

Please sign up or login with your details

Forgot password? Click here to reset