Cross-Batch Negative Sampling for Training Two-Tower Recommenders

10/28/2021
by   Jinpeng Wang, et al.
0

The two-tower architecture has been widely applied for learning item and user representations, which is important for large-scale recommender systems. Many two-tower models are trained using various in-batch negative sampling strategies, where the effects of such strategies inherently rely on the size of mini-batches. However, training two-tower models with a large batch size is inefficient, as it demands a large volume of memory for item and user contents and consumes a lot of time for feature encoding. Interestingly, we find that neural encoders can output relatively stable features for the same input after warming up in the training process. Based on such facts, we propose a simple yet effective sampling strategy called Cross-Batch Negative Sampling (CBNS), which takes advantage of the encoded item embeddings from recent mini-batches to boost the model training. Both theoretical analysis and empirical evaluations demonstrate the effectiveness and the efficiency of CBNS.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
12/14/2019

Cross-Batch Memory for Embedding Learning

Mining informative negative instances are of central importance to deep ...
research
05/30/2022

Cache-Augmented Inbatch Importance Resampling for Training Recommender Retriever

Recommender retrievers aim to rapidly retrieve a fraction of items from ...
research
06/19/2017

An Empirical Study of Mini-Batch Creation Strategies for Neural Machine Translation

Training of neural machine translation (NMT) models usually uses mini-ba...
research
05/05/2017

A comprehensive study of batch construction strategies for recurrent neural networks in MXNet

In this work we compare different batch construction methods for mini-ba...
research
09/14/2023

Turning Dross Into Gold Loss: is BERT4Rec really better than SASRec?

Recently sequential recommendations and next-item prediction task has be...
research
12/14/2018

An Empirical Model of Large-Batch Training

In an increasing number of domains it has been demonstrated that deep le...
research
03/02/2021

Mini-Batch Consistent Slot Set Encoder for Scalable Set Encoding

Most existing set encoding algorithms operate under the assumption that ...

Please sign up or login with your details

Forgot password? Click here to reset