Preference-grounded Token-level Guidance for Language Model Fine-tuning

06/01/2023
by   Shentao Yang, et al.
0

Aligning language models (LMs) with preferences is an important problem in natural language generation. A key challenge is that preferences are typically provided at the sequence level while LM training and generation both occur at the token level. There is, therefore, a granularity mismatch between the preference and the LM training losses, which may complicate the learning problem. In this paper, we address this issue by developing an alternate training process, where we iterate between grounding the sequence-level preference into token-level training guidance, and improving the LM with the learned guidance. For guidance learning, we design a framework that extends the pairwise-preference learning in imitation learning to both variable-length LM generation and utilizing the preference among multiple generations. For LM training, based on the amount of supervised data, we present two minimalist learning objectives that utilize the learned guidance. In experiments, our method performs competitively on two distinct representative LM tasks – discrete-prompt generation and text summarization.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/01/2021

A General Language Assistant as a Laboratory for Alignment

Given the broad capabilities of large language models, it should be poss...
research
12/20/2022

Controllable Text Generation with Language Constraints

We consider the task of text generation in language models with constrai...
research
06/08/2023

SequenceMatch: Imitation Learning for Autoregressive Sequence Modelling with Backtracking

In many domains, autoregressive models can achieve low log-likelihood on...
research
12/03/2022

CoP: Factual Inconsistency Detection by Controlling the Preference

Abstractive summarization is the process of generating a summary given a...
research
05/27/2022

Controllable Text Generation with Neurally-Decomposed Oracle

We propose a general and efficient framework to control auto-regressive ...
research
02/12/2020

Learning to Compare for Better Training and Evaluation of Open Domain Natural Language Generation Models

Automated evaluation of open domain natural language generation (NLG) mo...
research
05/24/2023

Analyzing Influential Factors in Human Preference Judgments via GPT-4

Pairwise human judgments are pivotal in guiding large language models (L...

Please sign up or login with your details

Forgot password? Click here to reset