BERT-kNN: Adding a kNN Search Component to Pretrained Language Models for Better QA

05/02/2020
by   Nora Kassner, et al.
0

Khandelwal et al. (2020) show that a k-nearest-neighbor (kNN) component improves language modeling performance. We use this idea for open domain question answering (QA). To improve the recall of facts stated in the training text, we combine BERT (Devlin et al., 2019) with a kNN search over a large corpus. Our contributions are as follows. i) We outperform BERT on cloze-style QA by large margins without any further training. ii) We show that BERT often identifies the correct response category (e.g., central European city), but only kNN recovers the factually correct answer (e.g., "Vienna").

READ FULL TEXT

page 1

page 2

page 3

page 4

research
11/08/2019

Negated LAMA: Birds cannot fly

Pretrained language models have achieved remarkable improvements in a br...
research
11/09/2019

BERT is Not a Knowledge Base (Yet): Factual Knowledge vs. Name-Based Reasoning in Unsupervised QA

The BERT language model (LM) (Devlin et al., 2019) is surprisingly good ...
research
11/17/2017

Learning to Organize Knowledge with N-Gram Machines

Deep neural networks (DNNs) had great success on NLP tasks such as langu...
research
02/14/2023

Exploring Category Structure with Contextual Language Models and Lexical Semantic Networks

Recent work on predicting category structure with distributional models,...
research
10/20/2020

Optimal Subarchitecture Extraction For BERT

We extract an optimal subset of architectural parameters for the BERT ar...
research
08/22/2019

Multi-passage BERT: A Globally Normalized BERT Model for Open-domain Question Answering

BERT model has been successfully applied to open-domain QA tasks. Howeve...
research
04/12/2021

Factual Probing Is [MASK]: Learning vs. Learning to Recall

Petroni et al. (2019) demonstrated that it is possible to retrieve world...

Please sign up or login with your details

Forgot password? Click here to reset