Effective and Efficient Training for Sequential Recommendation Using Cumulative Cross-Entropy Loss

01/03/2023
by   Fangyu Li, et al.
0

Increasing research interests focus on sequential recommender systems, aiming to model dynamic sequence representation precisely. However, the most commonly used loss function in state-of-the-art sequential recommendation models has essential limitations. To name a few, Bayesian Personalized Ranking (BPR) loss suffers the vanishing gradient problem from numerous negative sampling and predictionbiases; Binary Cross-Entropy (BCE) loss subjects to negative sampling numbers, thereby it is likely to ignore valuable negative examples and reduce the training efficiency; Cross-Entropy (CE) loss only focuses on the last timestamp of the training sequence, which causes low utilization of sequence information and results in inferior user sequence representation. To avoid these limitations, in this paper, we propose to calculate Cumulative Cross-Entropy (CCE) loss over the sequence. CCE is simple and direct, which enjoys the virtues of painless deployment, no negative sampling, and effective and efficient training. We conduct extensive experiments on five benchmark datasets to demonstrate the effectiveness and efficiency of CCE. The results show that employing CCE loss on three state-of-the-art models GRU4Rec, SASRec, and S3-Rec can reach 125.63 ranking NDCG@5, respectively. Using CCE, the performance curve of the models on the test data increases rapidly with the wall clock time, and is superior to that of other loss functions in almost the whole process of model training.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/14/2023

Turning Dross Into Gold Loss: is BERT4Rec really better than SASRec?

Recently sequential recommendations and next-item prediction task has be...
research
04/25/2022

Determinantal Point Process Likelihoods for Sequential Recommendation

Sequential recommendation is a popular task in academic research and clo...
research
08/14/2023

gSASRec: Reducing Overconfidence in Sequential Recommendation Trained with Negative Sampling

A large catalogue size is one of the central challenges in training reco...
research
02/25/2019

Improving Neural Response Diversity with Frequency-Aware Cross-Entropy Loss

Sequence-to-Sequence (Seq2Seq) models have achieved encouraging performa...
research
07/18/2019

On the relation between Loss Functions and T-Norms

Deep learning has been shown to achieve impressive results in several do...
research
10/29/2022

Reformulating van Rijsbergen's F_β metric for weighted binary cross-entropy

The separation of performance metrics from gradient based loss functions...
research
11/23/2016

Tunable Sensitivity to Large Errors in Neural Network Training

When humans learn a new concept, they might ignore examples that they ca...

Please sign up or login with your details

Forgot password? Click here to reset