Discourse-Aware Prompt Design for Text Generation

12/10/2021
by   Marjan Ghazvininejad, et al.
3

Current efficient fine-tuning methods (e.g., adapters, prefix-tuning, etc.) have optimized conditional text generation via training a small set of extra parameters of the neural language model, while freezing the rest for efficiency. While showing strong performance on some generation tasks, they don't generalize across all generation tasks. In this work, we show that prompt based conditional text generation can be improved with simple and efficient methods that simulate modeling the discourse structure of human written text. We introduce two key design choices: First we show that a higher-level discourse structure of human written text can be modelled with hierarchical blocking on prefix parameters that enable spanning different parts of the input and output text and yield more coherent output generations. Second, we propose sparse prefix tuning by introducing attention sparsity on the prefix parameters at different layers of the network and learn sparse transformations on the softmax-function, respectively. We find that sparse attention enables the prefix-tuning to better control of the input contents (salient facts) yielding more efficient tuning of the prefix-parameters. Experiments on a wide-variety of text generation tasks show that structured design of prefix parameters can achieve comparable results to fine-tuning all parameters while outperforming standard prefix-tuning on all generation tasks even in low-resource settings.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/17/2021

Stage-wise Fine-tuning for Graph-to-Text Generation

Graph-to-text generation has benefited from pre-trained language models ...
research
05/03/2022

Learning to Transfer Prompts for Text Generation

Pretrained language models (PLMs) have made remarkable progress in text ...
research
12/08/2020

Facts2Story: Controlling Text Generation by Key Facts

Recent advancements in self-attention neural network architectures have ...
research
08/09/2023

Emotion-Conditioned Text Generation through Automatic Prompt Optimization

Conditional natural language generation methods often require either exp...
research
05/10/2018

Discourse-Aware Neural Rewards for Coherent Text Generation

In this paper, we investigate the use of discourse-aware rewards with re...
research
06/17/2023

KEST: Kernel Distance Based Efficient Self-Training for Improving Controllable Text Generation

Self-training (ST) has come to fruition in language understanding tasks ...
research
04/17/2020

Rigid Formats Controlled Text Generation

Neural text generation has made tremendous progress in various tasks. On...

Please sign up or login with your details

Forgot password? Click here to reset