SamToNe: Improving Contrastive Loss for Dual Encoder Retrieval Models with Same Tower Negatives

06/05/2023
by   Fedor Moiseev, et al.
0

Dual encoders have been used for retrieval tasks and representation learning with good results. A standard way to train dual encoders is using a contrastive loss with in-batch negatives. In this work, we propose an improved contrastive learning objective by adding queries or documents from the same encoder towers to the negatives, for which we name it as "contrastive loss with SAMe TOwer NEgatives" (SamToNe). By evaluating on question answering retrieval benchmarks from MS MARCO and MultiReQA, and heterogenous zero-shot information retrieval benchmarks (BEIR), we demonstrate that SamToNe can effectively improve the retrieval quality for both symmetric and asymmetric dual encoders. By directly probing the embedding spaces of the two encoding towers via the t-SNE algorithm (van der Maaten and Hinton, 2008), we observe that SamToNe ensures the alignment between the embedding spaces from the two encoder towers. Based on the analysis of the embedding distance distributions of the top-1 retrieved results, we further explain the efficacy of the method from the perspective of regularisation.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/14/2022

Exploring Dual Encoder Architectures for Question Answering

Dual encoders have been used for question-answering (QA) and information...
research
11/09/2022

Distribution-Aligned Fine-Tuning for Efficient Neural Retrieval

Dual-encoder-based neural retrieval models achieve appreciable performan...
research
04/06/2023

Noise-Robust Dense Retrieval via Contrastive Alignment Post Training

The success of contextual word representations and advances in neural in...
research
01/18/2021

Scaling Deep Contrastive Learning Batch Size with Almost Constant Peak Memory Usage

Contrastive learning has been applied successfully to learn numerical ve...
research
05/04/2022

Analysing the Robustness of Dual Encoders for Dense Retrieval Against Misspellings

Dense retrieval is becoming one of the standard approaches for document ...
research
10/07/2021

Adversarial Retriever-Ranker for dense text retrieval

Current dense text retrieval models face two typical challenges. First, ...
research
04/28/2022

Keep the Caption Information: Preventing Shortcut Learning in Contrastive Image-Caption Retrieval

To train image-caption retrieval (ICR) methods, contrastive loss functio...

Please sign up or login with your details

Forgot password? Click here to reset