When to Fold'em: How to answer Unanswerable questions

05/01/2021
by   Marshall Ho, et al.
0

We present 3 different question-answering models trained on the SQuAD2.0 dataset – BIDAF, DocumentQA and ALBERT Retro-Reader – demonstrating the improvement of language models in the past three years. Through our research in fine-tuning pre-trained models for question-answering, we developed a novel approach capable of achieving a 2 training time. Our method of re-initializing select layers of a parameter-shared language model is simple yet empirically powerful.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/14/2020

Utilizing Bidirectional Encoder Representations from Transformers for Answer Selection

Pre-training a transformer-based model for the language modeling task in...
research
02/23/2023

Can Pre-trained Vision and Language Models Answer Visual Information-Seeking Questions?

Large language models have demonstrated an emergent capability in answer...
research
05/09/2023

Large Language Model Programs

In recent years, large pre-trained language models (LLMs) have demonstra...
research
02/10/2020

How Much Knowledge Can You Pack Into the Parameters of a Language Model?

It has recently been observed that neural language models trained on uns...
research
08/17/2022

ILLUME: Rationalizing Vision-Language Models by Interacting with their Jabber

Bootstrapping from pre-trained language models has been proven to be an ...
research
08/22/2023

Exploring the Effectiveness of GPT Models in Test-Taking: A Case Study of the Driver's License Knowledge Test

Large language models such as Open AI's Generative Pre-trained Transform...
research
04/20/2020

Fine-tuning Multi-hop Question Answering with Hierarchical Graph Network

In this paper, we present a two stage model for multi-hop question answe...

Please sign up or login with your details

Forgot password? Click here to reset