CREAK: A Dataset for Commonsense Reasoning over Entity Knowledge

09/03/2021
by   Yasumasa Onoe, et al.
0

Most benchmark datasets targeting commonsense reasoning focus on everyday scenarios: physical knowledge like knowing that you could fill a cup under a waterfall [Talmor et al., 2019], social knowledge like bumping into someone is awkward [Sap et al., 2019], and other generic situations. However, there is a rich space of commonsense inferences anchored to knowledge about specific entities: for example, deciding the truthfulness of a claim "Harry Potter can teach classes on how to fly on a broomstick." Can models learn to combine entity knowledge with commonsense reasoning in this fashion? We introduce CREAK, a testbed for commonsense reasoning about entity knowledge, bridging fact-checking about entities (Harry Potter is a wizard and is skilled at riding a broomstick) with commonsense inferences (if you're good at a skill you can teach others how to do it). Our dataset consists of 13k human-authored English claims about entities that are either true or false, in addition to a small contrast set. Crowdworkers can easily come up with these statements and human performance on the dataset is high (high 90s); we argue that models should be able to blend entity knowledge and commonsense reasoning to do well here. In our experiments, we focus on the closed-book setting and observe that a baseline model finetuned on existing fact verification benchmark struggles on CREAK. Training a model on CREAK improves accuracy by a substantial margin, but still falls short of human performance. Our benchmark provides a unique probe into natural language understanding models, testing both its ability to retrieve facts (e.g., who teaches at the University of Chicago?) and unstated commonsense knowledge (e.g., butlers do not yell at guests).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
10/01/2018

A Simple Machine Learning Method for Commonsense Reasoning? A Short Commentary on Trinh & Le (2018)

This is a short Commentary on Trinh & Le (2018) ("A Simple Method for Co...
research
05/10/2023

Decker: Double Check with Heterogeneous Knowledge for Commonsense Fact Verification

Commonsense fact verification, as a challenging branch of commonsense qu...
research
06/02/2021

COM2SENSE: A Commonsense Reasoning Benchmark with Complementary Sentences

Commonsense reasoning is intuitive for humans but has been a long-term c...
research
04/20/2021

Identify, Align, and Integrate: Matching Knowledge Graphs to Commonsense Reasoning Tasks

Integrating external knowledge into commonsense reasoning tasks has show...
research
05/22/2022

Commonsense Knowledge Salience Evaluation with a Benchmark Dataset in E-commerce

In e-commerce, the salience of commonsense knowledge (CSK) is beneficial...
research
05/21/2021

Fact-driven Logical Reasoning

Logical reasoning, which is closely related to human cognition, is of vi...
research
10/23/2022

ComFact: A Benchmark for Linking Contextual Commonsense Knowledge

Understanding rich narratives, such as dialogues and stories, often requ...

Please sign up or login with your details

Forgot password? Click here to reset