Show Your Work: Scratchpads for Intermediate Computation with Language Models

11/30/2021
by   Maxwell Nye, et al.
4

Large pre-trained language models perform remarkably well on tasks that can be done "in one pass", such as generating realistic text or synthesizing computer programs. However, they struggle with tasks that require unbounded multi-step computation, such as adding integers or executing programs. Surprisingly, we find that these same models are able to perform complex multi-step computations – even in the few-shot regime – when asked to perform the operation "step by step", showing the results of intermediate computations. In particular, we train transformers to perform multi-step computations by asking them to emit intermediate computation steps into a "scratchpad". On a series of increasingly complex tasks ranging from long addition to the execution of arbitrary programs, we show that scratchpads dramatically improve the ability of language models to perform multi-step computations.

READ FULL TEXT
research
11/22/2022

Program of Thoughts Prompting: Disentangling Computation from Reasoning for Numerical Reasoning Tasks

Recently, there has been significant progress in teaching language model...
research
03/16/2023

ART: Automatic multi-step reasoning and tool-use for large language models

Large language models (LLMs) can perform complex reasoning in few- and z...
research
05/19/2022

Selection-Inference: Exploiting Large Language Models for Interpretable Logical Reasoning

Large language models (LLMs) have been shown to be capable of impressive...
research
03/16/2022

Shepherd Pre-trained Language Models to Develop a Train of Thought: An Iterative Prompting Approach

While Pre-trained Language Models (PLMs) internalize a great amount of w...
research
04/07/2023

Why think step-by-step? Reasoning emerges from the locality of experience

Humans have a powerful and mysterious capacity to reason. By working thr...
research
08/09/2022

Limitations of Language Models in Arithmetic and Symbolic Induction

Recent work has shown that large pretrained Language Models (LMs) can no...
research
08/24/2018

Making Bubbling Practical

Bubbling is a run-time graph transformation studied for the execution of...

Please sign up or login with your details

Forgot password? Click here to reset