Hypothesis Search: Inductive Reasoning with Language Models

09/11/2023
by   Ruocheng Wang, et al.
0

Inductive reasoning is a core problem-solving capacity: humans can identify underlying principles from a few examples, which can then be robustly generalized to novel scenarios. Recent work has evaluated large language models (LLMs) on inductive reasoning tasks by directly prompting them yielding "in context learning." This can work well for straightforward inductive tasks, but performs very poorly on more complex tasks such as the Abstraction and Reasoning Corpus (ARC). In this work, we propose to improve the inductive reasoning ability of LLMs by generating explicit hypotheses at multiple levels of abstraction: we prompt the LLM to propose multiple abstract hypotheses about the problem, in natural language, then implement the natural language hypotheses as concrete Python programs. These programs can be directly verified by running on the observed examples and generalized to novel inputs. Because of the prohibitive cost of generation with state-of-the-art LLMs, we consider a middle step to filter the set of hypotheses that will be implemented into programs: we either ask the LLM to summarize into a smaller set of hypotheses, or ask human annotators to select a subset of the hypotheses. We verify our pipeline's effectiveness on the ARC visual inductive reasoning benchmark, its variant 1D-ARC, and string transformation dataset SyGuS. On a random 40-problem subset of ARC, our automated pipeline using LLM summaries achieves 27.5 accuracy, significantly outperforming the direct prompting baseline (accuracy of 12.5 candidates, the performance is boosted to 37.5 bound on the performance of our approach without filtering.) Our ablation studies show that abstract hypothesis generation and concrete program representations are both beneficial for LLMs to perform inductive reasoning tasks.

READ FULL TEXT

page 2

page 4

page 7

page 9

page 15

page 16

page 17

research
11/18/2022

PAL: Program-aided Language Models

Large language models (LLMs) have recently demonstrated an impressive ab...
research
12/21/2022

Language Models as Inductive Reasoners

Inductive reasoning is a core component of human intelligence. In the pa...
research
05/26/2023

LLMs and the Abstraction and Reasoning Corpus: Successes, Failures, and the Importance of Object-based Representations

Can a Large Language Model (LLM) solve simple abstract reasoning problem...
research
04/20/2023

Learning to Program with Natural Language

Large Language Models (LLMs) have shown remarkable performance in variou...
research
03/10/2021

Fast and flexible: Human program induction in abstract reasoning tasks

The Abstraction and Reasoning Corpus (ARC) is a challenging program indu...
research
10/18/2022

Graphs, Constraints, and Search for the Abstraction and Reasoning Corpus

The Abstraction and Reasoning Corpus (ARC) aims at benchmarking the perf...
research
05/30/2023

SheetCopilot: Bringing Software Productivity to the Next Level through Large Language Models

Computer end users have spent billions of hours completing daily tasks l...

Please sign up or login with your details

Forgot password? Click here to reset