ProtoQA: A Question Answering Dataset for Prototypical Common-Sense Reasoning

05/02/2020
by   Michael Boratko, et al.
0

Given questions regarding some prototypical situation – such as Name something that people usually do before they leave the house for work? – a human can easily answer them via acquired experiences. There can be multiple right answers for such questions with some more common for a situation than others. This paper introduces a new question answering dataset for training and evaluating common-sense reasoning capabilities of artificial intelligence systems in such prototypical situations. The training set is gathered from an existing set of questions played in a long-running international trivia game show – Family Feud. The hidden evaluation set is created by gathering answers for each question from 100 crowd-workers. We also propose an open-domain task where a model has to output a ranked list of answers, ideally covering all prototypical answers for a question. On evaluating our dataset with various competitive state-of-the-art models, we find there is a significant gap between the best model and human performance on a number of evaluation metrics.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/08/2019

AQuA: An Adversarially Authored Question-Answer Dataset for Common Sense

Commonsense reasoning is a critical AI capability, but it is difficult t...
research
08/27/2015

Using Thought-Provoking Children's Questions to Drive Artificial Intelligence Research

We propose to use thought-provoking children's questions (TPCQs), namely...
research
04/08/2019

CODAH: An Adversarially Authored Question-Answer Dataset for Common Sense

Commonsense reasoning is a critical AI capability, but it is difficult t...
research
06/24/2018

One-shot Learning for Question-Answering in Gaokao History Challenge

Answering questions from university admission exams (Gaokao in Chinese) ...
research
11/06/2019

A Spoken Dialogue System for Spatial Question Answering in a Physical Blocks World

The blocks world is a classic toy domain that has long been used to buil...
research
08/06/2020

Question and Answer Test-Train Overlap in Open-Domain Question Answering Datasets

Ideally Open-Domain Question Answering models should exhibit a number of...
research
05/25/2022

Reasoning over Logically Interacted Conditions for Question Answering

Some questions have multiple answers that are not equally correct, i.e. ...

Please sign up or login with your details

Forgot password? Click here to reset