Delaying Interaction Layers in Transformer-based Encoders for Efficient Open Domain Question Answering

10/16/2020
by   Wissam Siblini, et al.
0

Open Domain Question Answering (ODQA) on a large-scale corpus of documents (e.g. Wikipedia) is a key challenge in computer science. Although transformer-based language models such as Bert have shown on SQuAD the ability to surpass humans for extracting answers in small passages of text, they suffer from their high complexity when faced to a much larger search space. The most common way to tackle this problem is to add a preliminary Information Retrieval step to heavily filter the corpus and only keep the relevant passages. In this paper, we propose a more direct and complementary solution which consists in applying a generic change in the architecture of transformer-based models to delay the attention between subparts of the input and allow a more efficient management of computations. The resulting variants are competitive with the original models on the extractive task and allow, on the ODQA setting, a significant speedup and even a performance improvement in many cases.

READ FULL TEXT

page 2

page 3

research
09/26/2020

Techniques to Improve Q A Accuracy with Transformer-based models on Large Complex Documents

This paper discusses the effectiveness of various text processing techni...
research
02/05/2019

End-to-End Open-Domain Question Answering with BERTserini

We demonstrate an end-to-end question answering system that integrates B...
research
09/05/2023

Augmenting Black-box LLMs with Medical Textbooks for Clinical Question Answering

Large-scale language models (LLMs), such as ChatGPT, are capable of gene...
research
06/08/2023

Mapping the Challenges of HCI: An Application and Evaluation of ChatGPT and GPT-4 for Cost-Efficient Question Answering

Large language models (LLMs), such as ChatGPT and GPT-4, are gaining wid...
research
02/28/2020

DC-BERT: Decoupling Question and Document for Efficient Contextual Encoding

Recent studies on open-domain question answering have achieved prominent...
research
12/21/2022

Analyzing Semantic Faithfulness of Language Models via Input Intervention on Conversational Question Answering

Transformer-based language models have been shown to be highly effective...
research
05/14/2019

Multi-step Retriever-Reader Interaction for Scalable Open-domain Question Answering

This paper introduces a new framework for open-domain question answering...

Please sign up or login with your details

Forgot password? Click here to reset