DC-BERT: Decoupling Question and Document for Efficient Contextual Encoding

02/28/2020
by   Yuyu Zhang, et al.
0

Recent studies on open-domain question answering have achieved prominent performance improvement using pre-trained language models such as BERT. State-of-the-art approaches typically follow the "retrieve and read" pipeline and employ BERT-based reranker to filter retrieved documents before feeding them into the reader module. The BERT retriever takes as input the concatenation of question and each retrieved document. Despite the success of these approaches in terms of QA accuracy, due to the concatenation, they can barely handle high-throughput of incoming questions each with a large collection of retrieved documents. To address the efficiency problem, we propose DC-BERT, a decoupled contextual encoding framework that has dual BERT models: an online BERT which encodes the question only once, and an offline BERT which pre-encodes all the documents and caches their encodings. On SQuAD Open and Natural Questions Open datasets, DC-BERT achieves 10x speedup on document retrieval, while retaining most (about 98 compared to state-of-the-art approaches for open-domain question answering.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/16/2020

DDRQA: Dynamic Document Reranking for Open-domain Multi-hop Question Answering

Open-domain multi-hop question answering (QA) requires to retrieve multi...
research
07/21/2023

Generator-Retriever-Generator: A Novel Approach to Open-domain Question Answering

Open-domain question answering (QA) tasks usually require the retrieval ...
research
03/04/2020

A Study on Efficiency, Accuracy and Document Structure for Answer Sentence Selection

An essential task of most Question Answering (QA) systems is to re-rank ...
research
06/16/2021

A Neural Model for Joint Document and Snippet Ranking in Question Answering for Large Document Collections

Question answering (QA) systems for large document collections typically...
research
09/15/2021

Topic Transferable Table Question Answering

Weakly-supervised table question-answering(TableQA) models have achieved...
research
12/13/2021

Roof-BERT: Divide Understanding Labour and Join in Work

Recent work on enhancing BERT-based language representation models with ...
research
10/16/2020

Delaying Interaction Layers in Transformer-based Encoders for Efficient Open Domain Question Answering

Open Domain Question Answering (ODQA) on a large-scale corpus of documen...

Please sign up or login with your details

Forgot password? Click here to reset