FlowSeq: Non-Autoregressive Conditional Sequence Generation with Generative Flow

09/05/2019
by   Xuezhe Ma, et al.
0

Most sequence-to-sequence (seq2seq) models are autoregressive; they generate each token by conditioning on previously generated tokens. In contrast, non-autoregressive seq2seq models generate all tokens in one pass, which leads to increased efficiency through parallel processing on hardware such as GPUs. However, directly modeling the joint distribution of all tokens simultaneously is challenging, and even with increasingly complex model structures accuracy lags significantly behind autoregressive models. In this paper, we propose a simple, efficient, and effective model for non-autoregressive sequence generation using latent variable models. Specifically, we turn to generative flow, an elegant technique to model complex distributions using neural networks, and design several layers of flow tailored for modeling the conditional density of sequential latent variables. We evaluate this model on three neural machine translation (NMT) benchmark datasets, achieving comparable performance with state-of-the-art non-autoregressive NMT models and almost constant decoding time w.r.t the sequence length.

READ FULL TEXT
research
02/19/2018

Deterministic Non-Autoregressive Neural Sequence Modeling by Iterative Refinement

We propose a conditional non-autoregressive neural sequence model based ...
research
11/15/2015

Mixtures of Sparse Autoregressive Networks

We consider high-dimensional distribution estimation through autoregress...
research
06/17/2020

Towards Recurrent Autoregressive Flow Models

Stochastic processes generated by non-stationary distributions are diffi...
research
06/04/2019

KERMIT: Generative Insertion-Based Modeling for Sequences

We present KERMIT, a simple insertion-based approach to generative model...
research
05/17/2023

Variable Length Embeddings

In this work, we introduce a novel deep learning architecture, Variable ...
research
05/25/2022

Conditional set generation using Seq2seq models

Conditional set generation learns a mapping from an input sequence of to...
research
03/09/2018

Fast Decoding in Sequence Models using Discrete Latent Variables

Autoregressive sequence models based on deep neural networks, such as RN...

Please sign up or login with your details

Forgot password? Click here to reset