Strategic Reasoning with Language Models

05/30/2023
by   Kanishk Gandhi, et al.
0

Strategic reasoning enables agents to cooperate, communicate, and compete with other agents in diverse situations. Existing approaches to solving strategic games rely on extensive training, yielding strategies that do not generalize to new scenarios or games without retraining. Large Language Models (LLMs), with their ability to comprehend and generate complex, context-rich language, could prove powerful as tools for strategic gameplay. This paper introduces an approach that uses pretrained LLMs with few-shot chain-of-thought examples to enable strategic reasoning for AI agents. Our approach uses systematically generated demonstrations of reasoning about states, values, and beliefs to prompt the model. Using extensive variations of simple matrix games, we show that strategies that are derived based on systematically generated prompts generalize almost perfectly to new game structures, alternate objectives, and hidden information. Additionally, we demonstrate our approach can lead to human-like negotiation strategies in realistic scenarios without any extra training or fine-tuning. Our results highlight the ability of LLMs, guided by systematic reasoning demonstrations, to adapt and excel in diverse strategic scenarios.

READ FULL TEXT

page 17

page 19

page 20

research
09/12/2023

Strategic Behavior of Large Language Models: Game Structure vs. Contextual Framing

This paper investigates the strategic decision-making capabilities of th...
research
09/15/2023

Chain-of-Thought Reasoning is a Policy Improvement Operator

Large language models have astounded the world with fascinating new capa...
research
07/31/2023

Deception Abilities Emerged in Large Language Models

Large language models (LLMs) are currently at the forefront of intertwin...
research
02/11/2021

A Logic for Conditional Local Strategic Reasoning

We consider systems of rational agents who act and interact in pursuit o...
research
10/08/2021

Pick Your Battles: Interaction Graphs as Population-Level Objectives for Strategic Diversity

Strategic diversity is often essential in games: in multi-player games, ...
research
05/29/2023

Leveraging Training Data in Few-Shot Prompting for Numerical Reasoning

Chain-of-thought (CoT) prompting with large language models has proven e...
research
01/01/2023

Second Thoughts are Best: Learning to Re-Align With Human Values from Text Edits

We present Second Thought, a new learning paradigm that enables language...

Please sign up or login with your details

Forgot password? Click here to reset