Interpretable Time-Budget-Constrained Contextualization for Re-Ranking

02/04/2020
by   Sebastian Hofstätter, et al.
0

Search engines operate under a strict time constraint as a fast response is paramount to user satisfaction. Thus, neural re-ranking models have a limited time-budget to re-rank documents. Given the same amount of time, a faster re-ranking model can incorporate more documents than a less efficient one, leading to a higher effectiveness. To utilize this property, we propose TK (Transformer-Kernel): a neural re-ranking model for ad-hoc search using an efficient contextualization mechanism. TK employs a very small number of Transformer layers (up to three) to contextualize query and document word embeddings. To score individual term interactions, we use a document-length enhanced kernel-pooling, which enables users to gain insight into the model. TK offers an optimal ratio between effectiveness and efficiency: under realistic time constraints (max. 200 ms per query) TK achieves the highest effectiveness in comparison to BERT and other re-ranking models. We demonstrate this on three large-scale ranking collections: MSMARCO-Passage, MSMARCO-Document, and TREC CAR. In addition, to gain insight into TK, we perform a clustered query analysis of TK's results, highlighting its strengths and weaknesses on queries with different types of information need and we show how to interpret the cause of ranking differences of two documents by comparing their internal scores.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/03/2019

TU Wien @ TREC Deep Learning '19 – Simple Contextualization for Re-ranking

The usage of neural network models puts multiple objectives in conflict ...
research
04/18/2021

Anytime Ranking on Document-Ordered Indexes

Inverted indexes continue to be a mainstay of text search engines, allow...
research
07/16/2018

Repeatability Corner Cases in Document Ranking: The Impact of Score Ties

Document ranking experiments should be repeatable: running the same rank...
research
07/04/2022

Understanding Performance of Long-Document Ranking Models through Comprehensive Evaluation and Leaderboarding

We carry out a comprehensive evaluation of 13 recent models for ranking ...
research
06/20/2017

End-to-End Neural Ad-hoc Ranking with Kernel Pooling

This paper proposes K-NRM, a kernel based neural model for document rank...
research
04/30/2020

Query-level Early Exit for Additive Learning-to-Rank Ensembles

Search engine ranking pipelines are commonly based on large ensembles of...
research
11/21/2019

Separate and Attend in Personal Email Search

In personal email search, user queries often impose different requiremen...

Please sign up or login with your details

Forgot password? Click here to reset