ED2LM: Encoder-Decoder to Language Model for Faster Document Re-ranking Inference

04/25/2022
by   Kai Hui, et al.
3

State-of-the-art neural models typically encode document-query pairs using cross-attention for re-ranking. To this end, models generally utilize an encoder-only (like BERT) paradigm or an encoder-decoder (like T5) approach. These paradigms, however, are not without flaws, i.e., running the model on all query-document pairs at inference-time incurs a significant computational cost. This paper proposes a new training and inference paradigm for re-ranking. We propose to finetune a pretrained encoder-decoder model using in the form of document to query generation. Subsequently, we show that this encoder-decoder architecture can be decomposed into a decoder-only language model during inference. This results in significant inference time speedups since the decoder-only architecture only needs to learn to interpret static encoder embeddings during inference. Our experiments show that this new paradigm achieves results that are comparable to the more expensive cross-attention ranking approaches while being up to 6.8X faster. We believe this work paves the way for more efficient neural rankers that leverage large pretrained models.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/08/2023

Decoder-Only or Encoder-Decoder? Interpreting Language Model as a Regularized Encoder-Decoder

The sequence-to-sequence (seq2seq) task aims at generating the target se...
research
05/23/2023

NAIL: Lexical Retrieval Indices with Efficient Non-Autoregressive Decoders

Neural document rerankers are extremely effective in terms of accuracy. ...
research
12/15/2022

FiDO: Fusion-in-Decoder optimized for stronger performance and faster inference

Fusion-in-Decoder (FiD) is a powerful retrieval-augmented language model...
research
08/26/2018

Title-Guided Encoding for Keyphrase Generation

Keyphrase generation (KG) aims to generate a set of keyphrases given a d...
research
05/08/2023

PromptRank: Unsupervised Keyphrase Extraction Using Prompt

The keyphrase extraction task refers to the automatic selection of phras...
research
05/18/2020

Inflecting when there's no majority: Limitations of encoder-decoder neural networks as cognitive models for German plurals

Can artificial neural networks learn to represent inflectional morpholog...
research
08/09/2021

Not quite there yet: Combining analogical patterns and encoder-decoder networks for cognitively plausible inflection

The paper presents four models submitted to Part 2 of the SIGMORPHON 202...

Please sign up or login with your details

Forgot password? Click here to reset