Improving Transformer-Kernel Ranking Model Using Conformer and Query Term Independence

04/19/2021
by   Bhaskar Mitra, et al.
0

The Transformer-Kernel (TK) model has demonstrated strong reranking performance on the TREC Deep Learning benchmark – and can be considered to be an efficient (but slightly less effective) alternative to other Transformer-based architectures that employ (i) large-scale pretraining (high training cost), (ii) joint encoding of query and document (high inference cost), and (iii) larger number of Transformer layers (both high training and high inference costs). Since, a variant of the TK model – called TKL – has been developed that incorporates local self-attention to efficiently process longer input sequences in the context of document ranking. In this work, we propose a novel Conformer layer as an alternative approach to scale TK to longer input sequences. Furthermore, we incorporate query term independence and explicit term matching to extend the model to the full retrieval setting. We benchmark our models under the strictly blind evaluation setting of the TREC 2020 Deep Learning track and find that our proposed architecture changes lead to improved retrieval quality over TKL. Our best model also outperforms all non-neural runs ("trad") and two-thirds of the pretrained Transformer-based runs ("nnlm") on NDCG@10.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/20/2020

Conformer-Kernel with Query Term Independence for Document Retrieval

The Transformer-Kernel (TK) model has demonstrated strong reranking perf...
research
11/14/2020

Conformer-Kernel with Query Term Independence at TREC 2020 Deep Learning Track

We benchmark Conformer-Kernel models under the strict blind evaluation s...
research
05/11/2020

Local Self-Attention over Long Text for Efficient Document Retrieval

Neural networks, particularly Transformer-based architectures, have achi...
research
04/29/2020

Efficient Document Re-Ranking for Transformers by Precomputing Term Representations

Deep pretrained transformer networks are effective at various ranking ta...
research
12/03/2019

TU Wien @ TREC Deep Learning '19 – Simple Contextualization for Re-ranking

The usage of neural network models puts multiple objectives in conflict ...
research
05/22/2022

Dynamic Query Selection for Fast Visual Perceiver

Transformers have been matching deep convolutional networks for vision a...
research
09/16/2022

SQ-Swin: a Pretrained Siamese Quadratic Swin Transformer for Lettuce Browning Prediction

Packaged fresh-cut lettuce is widely consumed as a major component of ve...

Please sign up or login with your details

Forgot password? Click here to reset