Explanation Selection Using Unlabeled Data for In-Context Learning

02/09/2023
by   Xi Ye, et al.
0

Recent work has addressed textual reasoning tasks by prompting large language models with explanations via the chain-of-thought paradigm. However, subtly different explanations can yield widely varying downstream task accuracy, so explanations that have not been "tuned" for a task, such as off-the-shelf explanations written by nonexperts, may lead to mediocre performance. This paper tackles the problem of how to optimize explanation-infused prompts in a black-box fashion. We first generate sets of candidate explanations for each example in the prompt using a leave-one-out scheme. We then use a two-stage framework where we first evaluate explanations for each in-context example in isolation according to proxy metrics. Finally, we search over sets of explanations to find a set which yields high performance against a silver-labeled development set, drawing inspiration from recent work on bootstrapping language models on unlabeled data. Across four textual reasoning tasks spanning question answering, mathematical reasoning, and natural language inference, results show that our proxy metrics correlate with ground truth accuracy and our overall method can effectively improve prompts over crowdworker annotations and naive search strategies.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/20/2022

Learn to Explain: Multimodal Reasoning via Thought Chains for Science Question Answering

When answering a question, humans utilize the information available acro...
research
04/05/2022

CLEVR-X: A Visual Reasoning Dataset for Natural Language Explanations

Providing explanations in the context of Visual Question Answering (VQA)...
research
11/25/2022

Complementary Explanations for Effective In-Context Learning

Large language models (LLMs) have exhibited remarkable capabilities in l...
research
06/03/2020

From Probability to Consilience: How Explanatory Values Implement Bayesian Reasoning

Recent work in cognitive science has uncovered a diversity of explanator...
research
12/23/2021

Prolog-based agnostic explanation module for structured pattern classification

This paper presents a Prolog-based reasoning module to generate counterf...
research
12/08/2022

Harnessing the Power of Multi-Task Pretraining for Ground-Truth Level Natural Language Explanations

Natural language explanations promise to offer intuitively understandabl...
research
05/19/2023

OPT-R: Exploring the Role of Explanations in Finetuning and Prompting for Reasoning Skills of Large Language Models

In this paper, we conduct a thorough investigation into the reasoning ca...

Please sign up or login with your details

Forgot password? Click here to reset