Perhaps PTLMs Should Go to School – A Task to Assess Open Book and Closed Book QA

10/04/2021
by   Manuel R. Ciosici, et al.
0

Our goal is to deliver a new task and leaderboard to stimulate research on question answering and pre-trained language models (PTLMs) to understand a significant instructional document, e.g., an introductory college textbook or a manual. PTLMs have shown great success in many question-answering tasks, given significant supervised training, but much less so in zero-shot settings. We propose a new task that includes two college-level introductory texts in the social sciences (American Government 2e) and humanities (U.S. History), hundreds of true/false statements based on review questions written by the textbook authors, validation/development tests based on the first eight chapters of the textbooks, blind tests based on the remaining textbook chapters, and baseline results given state-of-the-art PTLMs. Since the questions are balanced, random performance should be  50 BoolQ achieves the same performance, suggesting that the textbook's content is not pre-represented in the PTLM. Taking the exam closed book, but having read the textbook (i.e., adding the textbook to T5's pre-training), yields at best minor improvement (56 textbook (or perhaps misunderstood the questions). Performance is better ( 60 when the exam is taken open-book (i.e., allowing the machine to automatically retrieve a paragraph and use it to answer the question).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/03/2021

Can Generative Pre-trained Language Models Serve as Knowledge Bases for Closed-book QA?

Recent work has investigated the interesting question using pre-trained ...
research
10/31/2022

Query Refinement Prompts for Closed-Book Long-Form Question Answering

Large language models (LLMs) have been shown to perform well in answerin...
research
11/22/2021

Zero-Shot Open-Book Question Answering

Open book question answering is a subset of question answering tasks whe...
research
12/31/2020

Studying Strategically: Learning to Mask for Closed-book QA

Closed-book question-answering (QA) is a challenging task that requires ...
research
10/13/2022

Closed-book Question Generation via Contrastive Learning

Question Generation (QG) is a fundamental NLP task for many downstream a...
research
04/15/2021

Designing a Minimal Retrieve-and-Read System for Open-Domain Question Answering

In open-domain question answering (QA), retrieve-and-read mechanism has ...
research
11/23/2022

Can Open-Domain QA Reader Utilize External Knowledge Efficiently like Humans?

Recent state-of-the-art open-domain QA models are typically based on a t...

Please sign up or login with your details

Forgot password? Click here to reset