Control, Generate, Augment: A Scalable Framework for Multi-Attribute Text Generation

04/30/2020
by   Giuseppe Russo, et al.
0

In this work, we present a text generation approach with multi-attribute control for data augmentation. We introduce CGA, a Variational Autoencoder architecture, to control, generate, and augment text. CGA is able to generate natural sentences with multiple controlled attributes by combining adversarial learning with a context-aware loss. The scalability of our approach is established through a single discriminator, independently of the number of attributes. As the main application of our work, we test the potential of this new model in a data augmentation use case. In a downstream NLP task, the sentences generated by our CGA model not only show significant improvements over a strong baseline, but also a classification performance very similar to real data. Furthermore, we are able to show high quality, diversity and attribute control in the generated sentences through a series of automatic and human assessments.

READ FULL TEXT

page 6

page 7

page 12

page 13

page 14

research
06/01/2023

Focused Prefix Tuning for Controllable Text Generation

In a controllable text generation dataset, there exist unannotated attri...
research
11/03/2018

Content preserving text generation with attribute controls

In this work, we address the problem of modifying textual attributes of ...
research
11/03/2020

Conditioned Text Generation with Transfer for Closed-Domain Dialogue Systems

Scarcity of training data for task-oriented dialogue systems is a well k...
research
02/23/2021

Controllable and Diverse Text Generation in E-commerce

In E-commerce, a key challenge in text generation is to find a good trad...
research
11/10/2019

Stylized Text Generation Using Wasserstein Autoencoders with a Mixture of Gaussian Prior

Wasserstein autoencoders are effective for text generation. They do not ...
research
07/06/2023

PREADD: Prefix-Adaptive Decoding for Controlled Text Generation

We propose Prefix-Adaptive Decoding (PREADD), a flexible method for cont...
research
09/08/2021

Smelting Gold and Silver for Improved Multilingual AMR-to-Text Generation

Recent work on multilingual AMR-to-text generation has exclusively focus...

Please sign up or login with your details

Forgot password? Click here to reset