Towards Teachable Reasoning Systems

04/27/2022
by   Bhavana Dalvi, et al.
0

Our goal is a teachable reasoning system for question-answering (QA), where a user can interact with faithful answer explanations, and correct errors so that the system improves over time. Our approach is three-fold: First, generated chains of reasoning show how answers are implied by the system's own internal beliefs. Second, users can interact with the explanations to identify erroneous model beliefs and provide corrections. Third, we augment the model with a dynamic memory of such corrections. Retrievals from memory are used as additional context for QA, to help avoid previous mistakes in similar new situations - a novel type of memory-based continuous learning. To our knowledge, this is the first system to generate chains that are both faithful (the answer follows from the reasoning) and truthful (the chain reflects the system's own beliefs, as ascertained by self-querying). In evaluation, users judge that a majority (65 follows from a set of facts - substantially better than a high-performance baseline. We also find that using simulated feedback, our system (called EntailmentWriter) continually improves with time, requiring feedback on only 25 examples). We observe a similar trend with real users. This suggests new opportunities for using language models in an interactive setting where users can inspect, debug, correct, and improve a system's performance over time.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/21/2022

Entailer: Answering Questions with Faithful and Truthful Chains of Reasoning

Our goal is a question-answering (QA) system that can show how its answe...
research
05/23/2023

Language Models with Rationality

While large language models (LLMs) are proficient at question-answering ...
research
04/06/2022

Using Interactive Feedback to Improve the Accuracy and Explainability of Question Answering Systems Post-Deployment

Most research on question answering focuses on the pre-deployment stage;...
research
10/31/2019

Do Multi-hop Readers Dream of Reasoning Chains?

General Question Answering (QA) systems over texts require the multi-hop...
research
04/16/2021

Enriching a Model's Notion of Belief using a Persistent Memory

Although pretrained language models (PTLMs) have been shown to contain s...
research
07/07/2021

Robustifying Multi-hop QA through Pseudo-Evidentiality Training

This paper studies the bias problem of multi-hop question answering mode...
research
09/29/2021

BeliefBank: Adding Memory to a Pre-Trained Language Model for a Systematic Notion of Belief

Although pretrained language models (PTLMs) contain significant amounts ...

Please sign up or login with your details

Forgot password? Click here to reset