Large Language Models are In-Context Semantic Reasoners rather than Symbolic Reasoners

05/24/2023
by   Xiaojuan Tang, et al.
0

The emergent few-shot reasoning capabilities of Large Language Models (LLMs) have excited the natural language and machine learning community over recent years. Despite of numerous successful applications, the underlying mechanism of such in-context capabilities still remains unclear. In this work, we hypothesize that the learned semantics of language tokens do the most heavy lifting during the reasoning process. Different from human's symbolic reasoning process, the semantic representations of LLMs could create strong connections among tokens, thus composing a superficial logical chain. To test our hypothesis, we decouple semantics from the language reasoning process and evaluate three kinds of reasoning abilities, i.e., deduction, induction and abduction. Our findings reveal that semantics play a vital role in LLMs' in-context reasoning – LLMs perform significantly better when semantics are consistent with commonsense but struggle to solve symbolic or counter-commonsense reasoning tasks by leveraging in-context new knowledge. The surprising observations question whether modern LLMs have mastered the inductive, deductive and abductive reasoning abilities as in human intelligence, and motivate research on unveiling the magic existing within the black-box LLMs. On the whole, our analysis provides a novel perspective on the role of semantics in developing and evaluating language models' reasoning abilities. Code is available at <https://github.com/XiaojuanTang/ICSR>.

READ FULL TEXT

page 20

page 21

page 22

page 23

page 24

page 25

page 26

page 27

research
10/06/2022

Language Models are Multilingual Chain-of-Thought Reasoners

We evaluate the reasoning abilities of large language models in multilin...
research
08/29/2023

When Do Program-of-Thoughts Work for Reasoning?

The reasoning capabilities of Large Language Models (LLMs) play a pivota...
research
08/09/2022

Limitations of Language Models in Arithmetic and Symbolic Induction

Recent work has shown that large pretrained Language Models (LMs) can no...
research
05/24/2023

EvEval: A Comprehensive Evaluation of Event Semantics for Large Language Models

Events serve as fundamental units of occurrence within various contexts....
research
08/03/2023

Reasoning in Large Language Models Through Symbolic Math Word Problems

Large language models (LLMs) have revolutionized NLP by solving downstre...
research
09/01/2023

Taken out of context: On measuring situational awareness in LLMs

We aim to better understand the emergence of `situational awareness' in ...
research
05/01/2023

Evaluating statistical language models as pragmatic reasoners

The relationship between communicated language and intended meaning is o...

Please sign up or login with your details

Forgot password? Click here to reset