How Different are Pre-trained Transformers for Text Ranking?

04/05/2022
by   David Rau, et al.
0

In recent years, large pre-trained transformers have led to substantial gains in performance over traditional retrieval models and feedback approaches. However, these results are primarily based on the MS Marco/TREC Deep Learning Track setup, with its very particular setup, and our understanding of why and how these models work better is fragmented at best. We analyze effective BERT-based cross-encoders versus traditional BM25 ranking for the passage retrieval task where the largest gains have been observed, and investigate two main questions. On the one hand, what is similar? To what extent does the neural ranker already encompass the capacity of traditional rankers? Is the gain in performance due to a better ranking of the same documents (prioritizing precision)? On the other hand, what is different? Can it retrieve effectively documents missed by traditional systems (prioritizing recall)? We discover substantial differences in the notion of relevance identifying strengths and weaknesses of BERT that may inspire research for future improvement. Our results contribute to our understanding of (black-box) neural rankers relative to (well-understood) traditional rankers, help understand the particular experimental setting of MS-Marco-based test collections.

READ FULL TEXT
research
05/05/2019

Investigating the Successes and Failures of BERT for Passage Re-Ranking

The bidirectional encoder representations from transformers (BERT) model...
research
08/15/2022

Continuous Active Learning Using Pretrained Transformers

Pre-trained and fine-tuned transformer models like BERT and T5 have impr...
research
01/20/2022

Transfer Learning Approaches for Building Cross-Language Dense Retrieval Models

The advent of transformer-based models such as BERT has led to the rise ...
research
03/14/2022

Can pre-trained Transformers be used in detecting complex sensitive sentences? – A Monsanto case study

Each and every organisation releases information in a variety of forms r...
research
10/01/2021

Building an Efficient and Effective Retrieval-based Dialogue System via Mutual Learning

Establishing retrieval-based dialogue systems that can select appropriat...
research
02/14/2022

DS4DH at TREC Health Misinformation 2021: Multi-Dimensional Ranking Models with Transfer Learning and Rank Fusion

This paper describes the work of the Data Science for Digital Health (DS...
research
08/31/2021

Shallow pooling for sparse labels

Recent years have seen enormous gains in core IR tasks, including docume...

Please sign up or login with your details

Forgot password? Click here to reset