Wacky Weights in Learned Sparse Representations and the Revenge of Score-at-a-Time Query Evaluation

10/22/2021
by   Joel Mackenzie, et al.
0

Recent advances in retrieval models based on learned sparse representations generated by transformers have led us to, once again, consider score-at-a-time query evaluation techniques for the top-k retrieval problem. Previous studies comparing document-at-a-time and score-at-a-time approaches have consistently found that the former approach yields lower mean query latency, although the latter approach has more predictable query latency. In our experiments with four different retrieval models that exploit representational learning with bags of words, we find that transformers generate "wacky weights" that appear to greatly reduce the opportunities for skipping and early exiting optimizations that lie at the core of standard document-at-a-time techniques. As a result, score-at-a-time approaches appear to be more competitive in terms of query evaluation latency than in previous studies. We find that, if an effectiveness loss of up to three percent can be tolerated, a score-at-a-time approach can yield substantial gains in mean query latency while at the same time dramatically reducing tail latency.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/24/2022

Faster Learned Sparse Retrieval with Guided Traversal

Neural information retrieval architectures based on transformers such as...
research
08/15/2022

Evaluating Dense Passage Retrieval using Transformers

Although representational retrieval models based on Transformers have be...
research
04/29/2020

Expansion via Prediction of Importance with Contextualization

The identification of relevance with little textual context is a primary...
research
07/31/2023

Lexically-Accelerated Dense Retrieval

Retrieval approaches that score documents based on learned dense vectors...
research
04/18/2021

Anytime Ranking on Document-Ordered Indexes

Inverted indexes continue to be a mainstay of text search engines, allow...
research
04/25/2023

Explain like I am BM25: Interpreting a Dense Model's Ranked-List with a Sparse Approximation

Neural retrieval models (NRMs) have been shown to outperform their stati...
research
05/02/2023

Optimizing Guided Traversal for Fast Learned Sparse Retrieval

Recent studies show that BM25-driven dynamic index skipping can greatly ...

Please sign up or login with your details

Forgot password? Click here to reset