Knowledge-driven Self-supervision for Zero-shot Commonsense Question Answering

11/07/2020
by   Kaixin Ma, et al.
9

Recent developments in pre-trained neural language modeling have led to leaps in accuracy on commonsense question-answering benchmarks. However, there is increasing concern that models overfit to specific tasks, without learning to utilize external knowledge or perform general semantic reasoning. In contrast, zero-shot evaluations have shown promise as a more robust measure of a model's general reasoning abilities. In this paper, we propose a novel neuro-symbolic framework for zero-shot question answering across commonsense tasks. Guided by a set of hypotheses, the framework studies how to transform various pre-existing knowledge resources into a form that is most effective for pre-training models. We vary the set of language models, training regimes, knowledge sources, and data generation strategies, and measure their impact across tasks. Extending on prior work, we devise and compare four constrained distractor-sampling strategies. We provide empirical results across five commonsense question-answering tasks with data generated from five external knowledge resources. We show that, while an individual knowledge graph is better suited for specific tasks, a global knowledge graph brings consistent gains across different tasks. In addition, both preserving the structure of the task as well as generating fair and informative questions help language models learn more effectively.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
05/10/2023

Multi-hop Commonsense Knowledge Injection Framework for Zero-Shot Commonsense Question Answering

Commonsense question answering (QA) research requires machines to answer...
research
04/11/2020

Unsupervised Commonsense Question Answering with Self-Talk

Natural language understanding involves reading between the lines with i...
research
06/07/2023

Knowledge-Augmented Language Model Prompting for Zero-Shot Knowledge Graph Question Answering

Large Language Models (LLMs) are capable of performing zero-shot closed-...
research
12/16/2021

DREAM: Uncovering Mental Models behind Language Models

To what extent do language models (LMs) build "mental models" of a scene...
research
11/02/2020

COSMO: Conditional SEQ2SEQ-based Mixture Model for Zero-Shot Commonsense Question Answering

Commonsense reasoning refers to the ability of evaluating a social situa...
research
03/23/2022

A Theoretically Grounded Benchmark for Evaluating Machine Commonsense

Programming machines with commonsense reasoning (CSR) abilities is a lon...
research
05/21/2022

An Empirical Investigation of Commonsense Self-Supervision with Knowledge Graphs

Self-supervision based on the information extracted from large knowledge...

Please sign up or login with your details

Forgot password? Click here to reset