Parallel Refinements for Lexically Constrained Text Generation with BART

09/26/2021
by   Xingwei He, et al.
0

Lexically constrained text generation aims to control the generated text by incorporating some pre-specified keywords into the output. Previous work injects lexical constraints into the output by controlling the decoding process or refining the candidate output iteratively, which tends to generate generic or ungrammatical sentences, and has high computational complexity. To address these challenges, we propose Constrained BART (CBART) for lexically constrained text generation. CBART leverages the pre-trained model BART and transfers part of the generation burden from the decoder to the encoder by decomposing this task into two sub-tasks, thereby improving the sentence quality. Concretely, we extend BART by adding a token-level classifier over the encoder, aiming at instructing the decoder where to replace and insert. Guided by the encoder, the decoder refines multiple tokens of the input in one step by inserting tokens before specific positions and re-predicting tokens with low confidence. To further reduce the inference latency, the decoder predicts all tokens in parallel. Experiment results on One-Billion-Word and Yelp show that CBART can generate plausible text with high quality and diversity while significantly accelerating inference.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/01/2020

POINTER: Constrained Text Generation via Insertion-based Generative Pre-training

Large-scale pre-trained language models, such as BERT and GPT-2, have ac...
research
11/15/2022

AutoTemplate: A Simple Recipe for Lexically Constrained Text Generation

Lexically constrained text generation is one of the constrained text gen...
research
09/13/2021

Show Me How To Revise: Improving Lexically Constrained Sentence Generation with XLNet

Lexically constrained sentence generation allows the incorporation of pr...
research
10/15/2020

Understanding Neural Abstractive Summarization Models via Uncertainty

An advantage of seq2seq abstractive summarization models is that they ge...
research
07/21/2021

Guided Generation of Cause and Effect

We present a conditional text generation framework that posits sententia...
research
05/24/2023

How To Control Text Simplification? An Empirical Study of Control Tokens for Meaning Preserving Controlled Simplification

Text simplification rewrites text to be more readable for a specific aud...
research
10/18/2018

Discourse Embellishment Using a Deep Encoder-Decoder Network

We suggest a new NLG task in the context of the discourse generation pip...

Please sign up or login with your details

Forgot password? Click here to reset