DuNST: Dual Noisy Self Training for Semi-Supervised Controllable Text Generation

12/16/2022
by   Yuxi Feng, et al.
0

Self-training (ST) has prospered again in language understanding by augmenting the fine-tuning of pre-trained language models when labeled data is insufficient. However, it remains challenging to incorporate ST into attribute-controllable language generation. Augmented by only self-generated pseudo text, generation models over-emphasize exploitation of the previously learned space, suffering from a constrained generalization boundary. We revisit ST and propose a novel method, DuNST to alleviate this problem. DuNST jointly models text generation and classification with a shared Variational AutoEncoder and corrupts the generated pseudo text by two kinds of flexible noise to disturb the space. In this way, our model could construct and utilize both pseudo text from given labels and pseudo labels from available unlabeled text, which are gradually refined during the ST process. We theoretically demonstrate that DuNST can be regarded as enhancing exploration towards the potential real text space, providing a guarantee of improved performance. Experiments on three controllable generation tasks show that DuNST could significantly boost control accuracy while maintaining comparable generation fluency and diversity against several strong baselines.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/17/2023

KEST: Kernel Distance Based Efficient Self-Training for Improving Controllable Text Generation

Self-training (ST) has come to fruition in language understanding tasks ...
research
06/06/2022

Curriculum-Based Self-Training Makes Better Few-Shot Learners for Data-to-Text Generation

Despite the success of text-to-text pre-trained models in various natura...
research
06/06/2023

Click: Controllable Text Generation with Sequence Likelihood Contrastive Learning

It has always been an important yet challenging problem to control langu...
research
03/24/2022

Mix and Match: Learning-free Controllable Text Generation using Energy Language Models

Recent work on controlled text generation has either required attribute-...
research
11/22/2022

Linear Interpolation In Parameter Space is Good Enough for Fine-Tuned Language Models

The simplest way to obtain continuous interpolation between two points i...
research
05/27/2022

Controllable Text Generation with Neurally-Decomposed Oracle

We propose a general and efficient framework to control auto-regressive ...
research
10/06/2022

FAST: Improving Controllability for Text Generation with Feedback Aware Self-Training

Controllable text generation systems often leverage control codes to dir...

Please sign up or login with your details

Forgot password? Click here to reset