Technical Report: Auxiliary Tuning and its Application to Conditional Text Generation

06/30/2020
by   Yoel Zeldes, et al.
0

We introduce a simple and efficient method, called Auxiliary Tuning, for adapting a pre-trained Language Model to a novel task; we demonstrate this approach on the task of conditional text generation. Our approach supplements the original pre-trained model with an auxiliary model that shifts the output distribution according to the target task. The auxiliary model is trained by adding its logits to the pre-trained model logits and maximizing the likelihood of the target task output. Our method imposes no constraints on the auxiliary architecture. In particular, the auxiliary model can ingest additional input relevant to the target task, independently from the pre-trained model's input. Furthermore, mixing the models at the logits level provides a natural probabilistic interpretation of the method. Our method achieved similar results to training from scratch for several different tasks, while using significantly fewer resources for training; we share a specific example of text generation conditioned on keywords.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/15/2021

Control Prefixes for Text Generation

Prompt learning methods adapt pre-trained language models to downstream ...
research
07/13/2020

Do You Have the Right Scissors? Tailoring Pre-trained Language Models via Monte-Carlo Methods

It has been a common approach to pre-train a language model on a large c...
research
10/05/2021

Exploring Conditional Text Generation for Aspect-Based Sentiment Analysis

Aspect-based sentiment analysis (ABSA) is an NLP task that entails proce...
research
12/09/2022

Fill in the Blank: Context-aware Automated Text Input Generation for Mobile GUI Testing

Automated GUI testing is widely used to help ensure the quality of mobil...
research
01/18/2023

Universal Neural-Cracking-Machines: Self-Configurable Password Models from Auxiliary Data

We develop the first universal password model – a password model that, o...
research
04/13/2022

GAP: A Graph-aware Language Model Framework for Knowledge Graph-to-Text Generation

Recent improvements in KG-to-text generation are due to additional auxil...
research
04/18/2021

Extract, Denoise, and Enforce: Evaluating and Predicting Lexical Constraints for Conditional Text Generation

Recently, pre-trained language models (PLMs) have dominated conditional ...

Please sign up or login with your details

Forgot password? Click here to reset