Think you have Solved Question Answering? Try ARC, the AI2 Reasoning Challenge

03/14/2018
by   Peter Clark, et al.
0

We present a new question set, text corpus, and baselines assembled to encourage AI research in advanced question answering. Together, these constitute the AI2 Reasoning Challenge (ARC), which requires far more powerful knowledge and reasoning than previous challenges such as SQuAD or SNLI. The ARC question set is partitioned into a Challenge Set and an Easy Set, where the Challenge Set contains only questions answered incorrectly by both a retrieval-based algorithm and a word co-occurence algorithm. The dataset contains only natural, grade-school science questions (authored for human tests), and is the largest public-domain set of this kind (7,787 questions). We test several baselines on the Challenge Set, including leading neural models from the SQuAD and SNLI tasks, and find that none are able to significantly outperform a random baseline, reflecting the difficult nature of this task. We are also releasing the ARC Corpus, a corpus of 14M science sentences relevant to the task, and implementations of the three neural baseline models tested. Can your model perform better? We pose ARC as a challenge to the community.

READ FULL TEXT

page 4

page 5

research
05/31/2018

KG^2: Learning to Reason Science Exam Questions with Contextual Knowledge Graph Embeddings

The AI2 Reasoning Challenge (ARC), a new benchmark dataset for question ...
research
06/01/2018

A Systematic Classification of Knowledge, Reasoning, and Context within the ARC Dataset

The recent work of Clark et al. introduces the AI2 Reasoning Challenge (...
research
04/26/2020

Challenge Closed-book Science Exam: A Meta-learning Based Question Answering System

Prior work in standardized science exams requires support from large tex...
research
09/04/2019

From 'F' to 'A' on the N.Y. Regents Science Exams: An Overview of the Aristo Project

AI has achieved remarkable mastery over games such as Chess, Go, and Pok...
research
10/02/2019

BookQA: Stories of Challenges and Opportunities

We present a system for answering questions based on the full text of bo...
research
08/31/2021

When Retriever-Reader Meets Scenario-Based Multiple-Choice Questions

Scenario-based question answering (SQA) requires retrieving and reading ...
research
06/06/2023

An Approach to Solving the Abstraction and Reasoning Corpus (ARC) Challenge

We utilise the power of Large Language Models (LLMs), in particular GPT4...

Please sign up or login with your details

Forgot password? Click here to reset