RARR: Researching and Revising What Language Models Say, Using Language Models

10/17/2022
by   Luyu Gao, et al.
0

Language models (LMs) now excel at many tasks such as few-shot learning, question answering, reasoning, and dialog. However, they sometimes generate unsupported or misleading content. A user cannot easily determine whether their outputs are trustworthy or not, because most LMs do not have any built-in mechanism for attribution to external evidence. To enable attribution while still preserving all the powerful advantages of recent generation models, we propose RARR (Retrofit Attribution using Research and Revision), a system that 1) automatically finds attribution for the output of any text generation model and 2) post-edits the output to fix unsupported content while preserving the original output as much as possible. When applied to the output of several state-of-the-art LMs on a diverse set of generation tasks, we find that RARR significantly improves attribution while otherwise preserving the original input to a much greater degree than previously explored edit models. Furthermore, the implementation of RARR requires only a handful of training examples, a large language model, and standard web search.

READ FULL TEXT

page 19

page 20

research
03/18/2022

Are You Robert or RoBERTa? Deceiving Online Authorship Attribution Models Using Neural Text Generators

Recently, there has been a rise in the development of powerful pre-train...
research
02/11/2023

Characterizing Attribution and Fluency Tradeoffs for Retrieval-Augmented Large Language Models

Despite recent progress, it has been difficult to prevent semantic hallu...
research
07/20/2023

The Extractive-Abstractive Axis: Measuring Content "Borrowing" in Generative Language Models

Generative language models produce highly abstractive outputs by design,...
research
05/10/2023

Automatic Evaluation of Attribution by Large Language Models

A recent focus of large language model (LLM) development, as exemplified...
research
08/18/2017

Assessing the Stylistic Properties of Neurally Generated Text in Authorship Attribution

Recent applications of neural language models have led to an increased i...
research
09/18/2023

Speaker attribution in German parliamentary debates with QLoRA-adapted large language models

The growing body of political texts opens up new opportunities for rich ...
research
06/21/2023

Feature Interactions Reveal Linguistic Structure in Language Models

We study feature interactions in the context of feature attribution meth...

Please sign up or login with your details

Forgot password? Click here to reset