Symbol tuning improves in-context learning in language models

05/15/2023
by   Jerry Wei, et al.
0

We present symbol tuning - finetuning language models on in-context input-label pairs where natural language labels (e.g., "positive/negative sentiment") are replaced with arbitrary symbols (e.g., "foo/bar"). Symbol tuning leverages the intuition that when a model cannot use instructions or natural language labels to figure out a task, it must instead do so by learning the input-label mappings. We experiment with symbol tuning across Flan-PaLM models up to 540B parameters and observe benefits across various settings. First, symbol tuning boosts performance on unseen in-context learning tasks and is much more robust to underspecified prompts, such as those without instructions or without natural language labels. Second, symbol-tuned models are much stronger at algorithmic reasoning tasks, with up to 18.2 Functions benchmark and up to 15.3 Concepts benchmark. Finally, symbol-tuned models show large improvements in following flipped-labels presented in-context, meaning that they are more capable of using in-context information to override prior semantic knowledge.

READ FULL TEXT

page 20

page 23

page 27

research
03/07/2023

Larger language models do in-context learning differently

We study how in-context learning (ICL) in language models is affected by...
research
04/26/2023

Exploring the Curious Case of Code Prompts

Recent work has shown that prompting language models with code-like repr...
research
08/14/2023

OctoPack: Instruction Tuning Code Large Language Models

Finetuning large language models (LLMs) on instructions leads to vast pe...
research
04/18/2021

Natural Instructions: Benchmarking Generalization to New Tasks from Natural Language Instructions

Can we enable NLP models to appropriately respond to instructional promp...
research
11/09/2022

Zero-Label Prompt Selection

Natural language prompts have been shown to facilitate cross-task genera...
research
09/03/2021

Symbol Emergence and The Solutions to Any Task

The following defines intent, an arbitrary task and its solutions, and t...
research
05/14/2019

Deep Residual Output Layers for Neural Language Generation

Many tasks, including language generation, benefit from learning the str...

Please sign up or login with your details

Forgot password? Click here to reset