Which Discriminator for Cooperative Text Generation?

04/25/2022
by   Antoine Chaffin, et al.
0

Language models generate texts by successively predicting probability distributions for next tokens given past ones. A growing field of interest tries to leverage external information in the decoding process so that the generated texts have desired properties, such as being more natural, non toxic, faithful, or having a specific writing style. A solution is to use a classifier at each generation step, resulting in a cooperative environment where the classifier guides the decoding of the language model distribution towards relevant texts for the task at hand. In this paper, we examine three families of (transformer-based) discriminators for this specific task of cooperative decoding: bidirectional, left-to-right and generative ones. We evaluate the pros and cons of these different types of discriminators for cooperative generation, exploring respective accuracy on classification tasks along with their impact on the resulting sample quality and computational performances. We also provide the code of a batched implementation of the powerful cooperative decoding strategy used for our experiments, the Monte Carlo Tree Search, working with each discriminator for Natural Language Generation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/18/2022

DisCup: Discriminator Cooperative Unlikelihood Prompt-tuning for Controllable Text Generation

Prompt learning with immensely large Casual Language Models (CLMs) has b...
research
04/27/2023

Controlled Text Generation with Natural Language Instructions

Large language models generate fluent texts and can follow natural langu...
research
08/10/2022

Controlling Perceived Emotion in Symbolic Music Generation with Monte Carlo Tree Search

This paper presents a new approach for controlling emotion in symbolic m...
research
12/20/2021

Spiral Language Modeling

In almost all text generation applications, word sequences are construct...
research
09/28/2021

Generating texts under constraint through discriminator-guided MCTS

Large pre-trained language models (LM) based on Transformers allow to ge...
research
05/06/2020

Token Manipulation Generative Adversarial Network for Text Generation

MaskGAN opens the query for the conditional language model by filling in...
research
12/12/2021

Towards More Efficient Insertion Transformer with Fractional Positional Encoding

Auto-regressive neural sequence models have been shown to be effective a...

Please sign up or login with your details

Forgot password? Click here to reset