Elaboration-Generating Commonsense Question Answering at Scale

09/02/2022
by   Wenya Wang, et al.
0

In question answering requiring common sense, language models (e.g., GPT-3) have been used to generate text expressing background knowledge that helps improve performance. Yet the cost of working with such models is very high; in this work, we finetune smaller language models to generate useful intermediate context, referred to here as elaborations. Our framework alternates between updating two language models – an elaboration generator and an answer predictor – allowing each to influence the other. Using less than 0.5 parameters of GPT-3, our model outperforms alternatives with similar sizes and closes the gap on GPT-3 on four commonsense question answering benchmarks. Human evaluations show that the quality of the generated elaborations is high.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/18/2019

Conversational AI : Open Domain Question Answering and Commonsense Reasoning

Our research is focused on making a human-like question answering system...
research
01/17/2022

Generalizable Neuro-symbolic Systems for Commonsense Question Answering

This chapter illustrates how suitable neuro-symbolic models for language...
research
01/04/2021

Benchmarking Knowledge-Enhanced Commonsense Question Answering via Knowledge-to-Text Transformation

A fundamental ability of humans is to utilize commonsense knowledge in l...
research
12/26/2021

ArT: All-round Thinker for Unsupervised Commonsense Question-Answering

Without labeled question-answer pairs for necessary training, unsupervis...
research
04/12/2020

Explaining Question Answering Models through Text Generation

Large pre-trained language models (LMs) have been shown to perform surpr...
research
03/27/2021

You Can Do Better! If You Elaborate the Reason When Making Prediction

Neural predictive models have achieved groundbreaking performance improv...
research
10/06/2020

Neural Mask Generator: Learning to Generate Adaptive Word Maskings for Language Model Adaptation

We propose a method to automatically generate a domain- and task-adaptiv...

Please sign up or login with your details

Forgot password? Click here to reset