Abductive Reasoning as Self-Supervision for Common Sense Question Answering

09/06/2019
by   Sathyanarayanan N. Aakur, et al.
0

Question answering has seen significant advances in recent times, especially with the introduction of increasingly bigger transformer-based models pre-trained on massive amounts of data. While achieving impressive results on many benchmarks, their performances appear to be proportional to the amount of training data available in the target domain. In this work, we explore the ability of current question-answering models to generalize - to both other domains as well as with restricted training data. We find that large amounts of training data are necessary, both for pre-training as well as fine-tuning to a task, for the models to perform well on the designated task. We introduce a novel abductive reasoning approach based on Grenander's Pattern Theory framework to provide self-supervised domain adaptation cues or "pseudo-labels," which can be used instead of expensive human annotations. The proposed self-supervised training regimen allows for effective domain adaptation without losing performance compared to fully supervised baselines. Extensive experiments on two publicly available benchmarks show the efficacy of the proposed approach. We show that neural networks models trained using self-labeled data can retain up to 75% of the performance of models trained on large amounts of human-annotated training data. Code and evaluation data will be made available publicly upon acceptance.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
06/17/2021

An Evaluation of Self-Supervised Pre-Training for Skin-Lesion Analysis

Self-supervised pre-training appears as an advantageous alternative to s...
research
11/13/2019

Neural Duplicate Question Detection without Labeled Training Data

Supervised training of neural models to duplicate question detection in ...
research
11/06/2019

Unsupervised Domain Adaptation of Contextual Embeddings for Low-Resource Duplicate Question Detection

Answering questions is a primary goal of many conversational systems or ...
research
03/31/2022

Domain Adaptation for Sparse-Data Settings: What Do We Gain by Not Using Bert?

The practical success of much of NLP depends on the availability of trai...
research
09/16/2020

Knowledge Guided Learning: Towards Open Domain Egocentric Action Recognition with Zero Supervision

Advances in deep learning have enabled the development of models that ha...
research
02/26/2021

DOCENT: Learning Self-Supervised Entity Representations from Large Document Collections

This paper explores learning rich self-supervised entity representations...

Please sign up or login with your details

Forgot password? Click here to reset