Generative Pre-training for Paraphrase Generation by Representing and Predicting Spans in Exemplars

11/29/2020
by   Tien-Cuong Bui, et al.
0

Paraphrase generation is a long-standing problem and serves an essential role in many natural language processing problems. Despite some encouraging results, recent methods either confront the problem of favoring generic utterance or need to retrain the model from scratch for each new dataset. This paper presents a novel approach to paraphrasing sentences, extended from the GPT-2 model. We develop a template masking technique, named first-order masking, to masked out irrelevant words in exemplars utilizing POS taggers. So that, the paraphrasing task is changed to predicting spans in masked templates. Our proposed approach outperforms competitive baselines, especially in the semantic preservation aspect. To prevent the model from being biased towards a given template, we introduce a technique, referred to as second-order masking, which utilizes Bernoulli distribution to control the visibility of the first-order-masked template's tokens. Moreover, this technique allows the model to provide various paraphrased sentences in testing by adjusting the second-order-masking level. For scale-up objectives, we compare the performance of two alternatives template-selection methods, which shows that they were equivalent in preserving semantic information.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/19/2022

Improving Aspect Sentiment Quad Prediction via Template-Order Data Augmentation

Recently, aspect sentiment quad prediction (ASQP) has become a popular t...
research
04/30/2020

Few-Shot Natural Language Generation by Rewriting Templates

Virtual assistants such as Google Assistant, Alexa and Siri enable users...
research
07/29/2019

ERNIE 2.0: A Continual Pre-training Framework for Language Understanding

Recently, pre-trained models have achieved state-of-the-art results in v...
research
04/19/2021

Natural Language Generation Using Link Grammar for General Conversational Intelligence

Many current artificial general intelligence (AGI) and natural language ...
research
09/28/2021

Template-free Prompt Tuning for Few-shot NER

Prompt-based methods have been successfully applied in sentence-level fe...
research
05/23/2023

TaDSE: Template-aware Dialogue Sentence Embeddings

Learning high quality sentence embeddings from dialogues has drawn incre...
research
04/21/2020

Vector Quantized Contrastive Predictive Coding for Template-based Music Generation

In this work, we propose a flexible method for generating variations of ...

Please sign up or login with your details

Forgot password? Click here to reset