Disentangling Reasoning Capabilities from Language Models with Compositional Reasoning Transformers

10/20/2022
by   Wanjun Zhong, et al.
0

This paper presents ReasonFormer, a unified reasoning framework for mirroring the modular and compositional reasoning process of humans in complex decision making. Inspired by dual-process theory in cognitive science, the representation module (automatic thinking) and reasoning modules (controlled thinking) are disentangled to capture different levels of cognition. Upon the top of the representation module, the pre-trained reasoning modules are modular and expertise in specific and fundamental reasoning skills (e.g., logic, simple QA, etc). To mimic the controlled compositional thinking process, different reasoning modules are dynamically activated and composed in both parallel and cascaded manners to control what reasoning skills are activated and how deep the reasoning process will be reached to solve the current problems. The unified reasoning framework solves multiple tasks with a single model,and is trained and inferred in an end-to-end manner. Evaluated on 11 datasets requiring different reasoning skills and complexity, ReasonFormer demonstrates substantial performance boosts, revealing the compositional reasoning ability. Few-shot experiments exhibit better generalization ability by learning to compose pre-trained skills for new tasks with limited data,and decoupling the representation module and the reasoning modules. Further analysis shows the modularity of reasoning modules as different tasks activate distinct reasoning skills at different reasoning depths.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/16/2022

ALERT: Adapting Language Models to Reasoning Tasks

Current large language models can perform reasonably well on complex tas...
research
09/14/2023

Dynamic MOdularized Reasoning for Compositional Structured Explanation Generation

Despite the success of neural models in solving reasoning tasks, their c...
research
08/01/2023

Skills-in-Context Prompting: Unlocking Compositionality in Large Language Models

We consider the problem of eliciting compositional generalization capabi...
research
07/23/2018

Explainable Neural Computation via Stack Neural Module Networks

In complex inferential tasks like question answering, machine learning m...
research
10/08/2019

Meta Module Network for Compositional Visual Reasoning

There are two main lines of research on visual reasoning: neural module ...
research
02/01/2022

Planner-Reasoner Inside a Multi-task Reasoning Agent

We consider the problem of multi-task reasoning (MTR), where an agent ca...
research
10/06/2022

Teaching Neural Module Networks to Do Arithmetic

Answering complex questions that require multi-step multi-type reasoning...

Please sign up or login with your details

Forgot password? Click here to reset