An Invariant Learning Characterization of Controlled Text Generation

05/31/2023
by   Carolina Zheng, et al.
0

Controlled generation refers to the problem of creating text that contains stylistic or semantic attributes of interest. Many approaches reduce this problem to training a predictor of the desired attribute. For example, researchers hoping to deploy a large language model to produce non-toxic content may use a toxicity classifier to filter generated text. In practice, the generated text to classify, which is determined by user prompts, may come from a wide range of distributions. In this paper, we show that the performance of controlled generation may be poor if the distributions of text in response to user prompts differ from the distribution the predictor was trained on. To address this problem, we cast controlled generation under distribution shift as an invariant learning problem: the most effective predictor should be invariant across multiple text environments. We then discuss a natural solution that arises from this characterization and propose heuristics for selecting natural environments. We study this characterization and the proposed method empirically using both synthetic and real data. Experiments demonstrate both the challenge of distribution shift in controlled generation and the potential of invariance methods in this setting.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/12/2021

FUDGE: Controlled Text Generation With Future Discriminators

We propose Future Discriminators for Generation (FUDGE), a flexible and ...
research
07/05/2021

Causally Invariant Predictor with Shift-Robustness

This paper proposes an invariant causal predictor that is robust to dist...
research
05/07/2021

On-the-Fly Controlled Text Generation with Experts and Anti-Experts

Despite recent advances in natural language generation, it remains chall...
research
12/13/2021

Controlled Cue Generation for Play Scripts

In this paper, we use a large-scale play scripts dataset to propose the ...
research
10/14/2022

PCFG-based Natural Language Interface Improves Generalization for Controlled Text Generation

Existing work on controlled text generation (CTG) assumes a control inte...
research
06/16/2021

Automatic Construction of Evaluation Suites for Natural Language Generation Datasets

Machine learning approaches applied to NLP are often evaluated by summar...
research
03/25/2020

Heavy-tailed Representations, Text Polarity Classification Data Augmentation

The dominant approaches to text representation in natural language rely ...

Please sign up or login with your details

Forgot password? Click here to reset