Fast Passage Re-ranking with Contextualized Exact Term Matching and Efficient Passage Expansion

08/19/2021
by   Shengyao Zhuang, et al.
0

BERT-based information retrieval models are expensive, in both time (query latency) and computational resources (energy, hardware cost), making many of these models impractical especially under resource constraints. The reliance on a query encoder that only performs tokenization and on the pre-processing of passage representations at indexing, has allowed the recently proposed TILDE method to overcome the high query latency issue typical of BERT-based models. This however is at the expense of a lower effectiveness compared to other BERT-based re-rankers and dense retrievers. In addition, the original TILDE method is characterised by indexes with a very high memory footprint, as it expands each passage into the size of the BERT vocabulary. In this paper, we propose TILDEv2, a new model that stems from the original TILDE but that addresses its limitations. TILDEv2 relies on contextualized exact term matching with expanded passages. This requires to only store in the index the score of tokens that appear in the expanded passages (rather than all the vocabulary), thus producing indexes that are 99 mechanism also improves ranking effectiveness by 24 query latency. This makes TILDEv2 the state-of-the-art passage re-ranking method for CPU-only environments, capable of maintaining query latency below 100ms on commodity hardware.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/29/2020

Expansion via Prediction of Importance with Contextualization

The identification of relevance with little textual context is a primary...
research
07/15/2020

Deep Reinforced Query Reformulation for Information Retrieval

Query reformulations have long been a key mechanism to alleviate the voc...
research
10/11/2022

On the Interpolation of Contextualized Term-based Ranking with BM25 for Query-by-Example Retrieval

Term-based ranking with pre-trained transformer-based language models ha...
research
04/24/2022

Faster Learned Sparse Retrieval with Guided Traversal

Neural information retrieval architectures based on transformers such as...
research
09/22/2021

Predicting Efficiency/Effectiveness Trade-offs for Dense vs. Sparse Retrieval Strategy Selection

Over the last few years, contextualized pre-trained transformer models s...
research
01/23/2023

Injecting the BM25 Score as Text Improves BERT-Based Re-rankers

In this paper we propose a novel approach for combining first-stage lexi...

Please sign up or login with your details

Forgot password? Click here to reset