No Parameter Left Behind: How Distillation and Model Size Affect Zero-Shot Retrieval

06/06/2022
by   Guilherme Moraes Rosa, et al.
0

Recent work has shown that small distilled language models are strong competitors to models that are orders of magnitude larger and slower in a wide range of information retrieval tasks. This has made distilled and dense models, due to latency constraints, the go-to choice for deployment in real-world retrieval applications. In this work, we question this practice by showing that the number of parameters and early query-document interaction play a significant role in the generalization ability of retrieval models. Our experiments show that increasing model size results in marginal gains on in-domain test sets, but much larger gains in new domains never seen during fine-tuning. Furthermore, we show that rerankers largely outperform dense ones of similar size in several tasks. Our largest reranker reaches the state of the art in 12 of the 18 datasets of the Benchmark-IR (BEIR) and surpasses the previous state of the art by 3 average points. Finally, we confirm that in-domain effectiveness is not a good indicator of zero-shot effectiveness. Code is available at https://github.com/guilhermemr04/scaling-zero-shot-retrieval.git

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/12/2022

In Defense of Cross-Encoders for Zero-Shot Retrieval

Bi-encoders and cross-encoders are widely used in many state-of-the-art ...
research
04/17/2021

BEIR: A Heterogenous Benchmark for Zero-shot Evaluation of Information Retrieval Models

Neural IR models have often been studied in homogeneous and narrow setti...
research
02/23/2023

On the Generalization Ability of Retrieval-Enhanced Transformers

Recent work on the Retrieval-Enhanced Transformer (RETRO) model has show...
research
07/08/2022

An Efficiency Study for SPLADE Models

Latency and efficiency issues are often overlooked when evaluating IR mo...
research
03/23/2023

Modular Retrieval for Generalization and Interpretation

New retrieval tasks have always been emerging, thus urging the developme...
research
08/16/2023

RSpell: Retrieval-augmented Framework for Domain Adaptive Chinese Spelling Check

Chinese Spelling Check (CSC) refers to the detection and correction of s...
research
08/24/2022

DPTDR: Deep Prompt Tuning for Dense Passage Retrieval

Deep prompt tuning (DPT) has gained great success in most natural langua...

Please sign up or login with your details

Forgot password? Click here to reset