EARL: Speedup Transformer-based Rankers with Pre-computed Representation

04/28/2020
by   Luyu Gao, et al.
0

Recent innovations in Transformer-based ranking models have advanced the state-of-the-art in information retrieval. However, their performance gains come at a steep computational cost. This paper presents a novel Embed Ahead Rank Later (EARL) framework, which speeds-up Transformer-based rankers by pre-computing representations and keeping online computation shallow. EARL dis-entangles the attention in a typical Transformer-based ranker into three asynchronous tasks and assign each to a dedicated Transformer: query understanding, document understanding, and relevance judging. With such a ranking framework, query and document token representations can be offline computed and reused. We also propose a new judger transformer block that keeps online relevance judging light and shallow. Our experiments demonstrate that EARL can be as effective as previous state-of-the-art BERT rankers in accuracy while substantially faster in evaluation time.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/29/2022

Compact Token Representations with Contextual Quantization for Efficient Document Re-ranking

Transformer based re-ranking models can achieve high search relevance th...
research
04/29/2020

Efficient Document Re-Ranking for Transformers by Precomputing Term Representations

Deep pretrained transformer networks are effective at various ranking ta...
research
03/11/2021

Composite Re-Ranking for Efficient Document Search with BERT

Although considerable efforts have been devoted to transformer-based ran...
research
10/11/2022

On the Interpolation of Contextualized Term-based Ranking with BM25 for Query-by-Example Retrieval

Term-based ranking with pre-trained transformer-based language models ha...
research
01/20/2021

PGT: Pseudo Relevance Feedback Using a Graph-Based Transformer

Most research on pseudo relevance feedback (PRF) has been done in vector...
research
10/03/2021

SDR: Efficient Neural Re-ranking using Succinct Document Representation

BERT based ranking models have achieved superior performance on various ...
research
04/26/2020

Choppy: Cut Transformer For Ranked List Truncation

Work in information retrieval has traditionally focused on ranking and r...

Please sign up or login with your details

Forgot password? Click here to reset