DINOISER: Diffused Conditional Sequence Learning by Manipulating Noises

02/20/2023
by   Jiasheng Ye, et al.
0

While diffusion models have achieved great success in generating continuous signals such as images and audio, it remains elusive for diffusion models in learning discrete sequence data like natural languages. Although recent advances circumvent this challenge of discreteness by embedding discrete tokens as continuous surrogates, they still fall short of satisfactory generation quality. To understand this, we first dive deep into the denoised training protocol of diffusion-based sequence generative models and determine their three severe problems, i.e., 1) failing to learn, 2) lack of scalability, and 3) neglecting source conditions. We argue that these problems can be boiled down to the pitfall of the not completely eliminated discreteness in the embedding space, and the scale of noises is decisive herein. In this paper, we introduce DINOISER to facilitate diffusion models for sequence generation by manipulating noises. We propose to adaptively determine the range of sampled noise scales for counter-discreteness training; and encourage the proposed diffused sequence learner to leverage source conditions with amplified noise scales during inference. Experiments show that DINOISER enables consistent improvement over the baselines of previous diffusion-based sequence generative models on several conditional sequence modeling benchmarks thanks to both effective training and inference strategies. Analyses further verify that DINOISER can make better use of source conditions to govern its generative process.

READ FULL TEXT
research
10/17/2022

DiffuSeq: Sequence to Sequence Text Generation with Diffusion Models

Recently, diffusion models have emerged as a new paradigm for generative...
research
05/18/2023

Dirichlet Diffusion Score Model for Biological Sequence Generation

Designing biological sequences is an important challenge that requires s...
research
12/20/2022

SeqDiffuSeq: Text Diffusion with Encoder-Decoder Transformers

Diffusion model, a new generative modelling paradigm, has achieved great...
research
02/10/2022

Conditional Diffusion Probabilistic Model for Speech Enhancement

Speech enhancement is a critical component of many user-oriented audio a...
research
03/08/2023

Diffusing Gaussian Mixtures for Generating Categorical Data

Learning a categorical distribution comes with its own set of challenges...
research
04/11/2023

Diffusion Recommender Model

Generative models such as Generative Adversarial Networks (GANs) and Var...
research
04/25/2023

CoDi: Co-evolving Contrastive Diffusion Models for Mixed-type Tabular Synthesis

With growing attention to tabular data these days, the attempt to apply ...

Please sign up or login with your details

Forgot password? Click here to reset