Mixture Content Selection for Diverse Sequence Generation

09/04/2019
by   Jaemin Cho, et al.
0

Generating diverse sequences is important in many NLP applications such as question generation or summarization that exhibit semantically one-to-many relationships between source and the target sequences. We present a method to explicitly separate diversification from generation using a general plug-and-play module (called SELECTOR) that wraps around and guides an existing encoder-decoder model. The diversification stage uses a mixture of experts to sample different binary masks on the source sequence for diverse content selection. The generation stage uses a standard encoder-decoder model given each selected content from the source sequence. Due to the non-differentiable nature of discrete sampling and the lack of ground truth labels for binary mask, we leverage a proxy for ground truth mask and adopt stochastic hard-EM for training. In question generation (SQuAD) and abstractive summarization (CNN-DM), our method demonstrates significant improvements in accuracy, diversity and training efficiency, including state-of-the-art top-1 accuracy in both datasets, 6 state of the art model. Our code is publicly available at https://github.com/clovaai/FocusSeq2Seq.

READ FULL TEXT
research
08/28/2018

Guided Neural Language Generation for Abstractive Summarization using Abstract Meaning Representation

Recent work on abstractive summarization has made progress with neural e...
research
02/25/2019

Pretraining-Based Natural Language Generation for Text Summarization

In this paper, we propose a novel pretraining-based encoder-decoder fram...
research
12/01/2019

End to End Trainable Active Contours via Differentiable Rendering

We present an image segmentation method that iteratively evolves a polyg...
research
06/15/2019

Mask Based Unsupervised Content Transfer

We consider the problem of translating, in an unsupervised manner, betwe...
research
01/27/2021

Scheduled Sampling in Vision-Language Pretraining with Decoupled Encoder-Decoder Network

Despite having impressive vision-language (VL) pretraining with BERT-bas...
research
06/05/2023

Bootstrapped Training of Score-Conditioned Generator for Offline Design of Biological Sequences

We study the problem of optimizing biological sequences, e.g., proteins,...
research
04/23/2017

Deep Keyphrase Generation

Keyphrase provides highly-summative information that can be effectively ...

Please sign up or login with your details

Forgot password? Click here to reset