LLMs and the Abstraction and Reasoning Corpus: Successes, Failures, and the Importance of Object-based Representations

05/26/2023
by   Yudong Xu, et al.
0

Can a Large Language Model (LLM) solve simple abstract reasoning problems? We explore this broad question through a systematic analysis of GPT on the Abstraction and Reasoning Corpus (ARC), a representative benchmark of abstract reasoning ability from limited examples in which solutions require some "core knowledge" of concepts such as objects, goal states, counting, and basic geometry. GPT-4 solves only 13/50 of the most straightforward ARC tasks when using textual encodings for their two-dimensional input-output grids. Our failure analysis reveals that GPT-4's capacity to identify objects and reason about them is significantly influenced by the sequential nature of the text that represents an object within a text encoding of a task. To test this hypothesis, we design a new benchmark, the 1D-ARC, which consists of one-dimensional (array-like) tasks that are more conducive to GPT-based reasoning, and where it indeed performs better than on the (2D) ARC. To alleviate this issue, we propose an object-based representation that is obtained through an external tool, resulting in nearly doubling the performance on solved ARC tasks and near-perfect scores on the easier 1D-ARC. Although the state-of-the-art GPT-4 is unable to "reason" perfectly within non-language domains such as the 1D-ARC or a simple ARC subset, our study reveals that the use of object-based representations can significantly improve its reasoning ability. Visualizations, GPT logs, and data are available at https://khalil-research.github.io/LLM4ARC.

READ FULL TEXT

page 2

page 3

page 4

page 6

page 12

page 13

page 17

research
06/04/2023

Systematic Visual Reasoning through Object-Centric Relational Abstraction

Human visual reasoning is characterized by an ability to identify abstra...
research
09/11/2023

Hypothesis Search: Inductive Reasoning with Language Models

Inductive reasoning is a core problem-solving capacity: humans can ident...
research
06/06/2023

An Approach to Solving the Abstraction and Reasoning Corpus (ARC) Challenge

We utilise the power of Large Language Models (LLMs), in particular GPT4...
research
10/18/2022

Graphs, Constraints, and Search for the Abstraction and Reasoning Corpus

The Abstraction and Reasoning Corpus (ARC) aims at benchmarking the perf...
research
01/05/2022

Does entity abstraction help generative Transformers reason?

Pre-trained language models (LMs) often struggle to reason logically or ...
research
11/29/2018

Learning to Reason with Third-Order Tensor Products

We combine Recurrent Neural Networks with Tensor Product Representations...
research
09/20/2022

FACT: Learning Governing Abstractions Behind Integer Sequences

Integer sequences are of central importance to the modeling of concepts ...

Please sign up or login with your details

Forgot password? Click here to reset