A Systematic Review and Replicability Study of BERT4Rec for Sequential Recommendation

07/15/2022
by   Aleksandr Petrov, et al.
0

BERT4Rec is an effective model for sequential recommendation based on the Transformer architecture. In the original publication, BERT4Rec claimed superiority over other available sequential recommendation approaches (e.g. SASRec), and it is now frequently being used as a state-of-the art baseline for sequential recommendations. However, not all subsequent publications confirmed this result and proposed other models that were shown to outperform BERT4Rec in effectiveness. In this paper we systematically review all publications that compare BERT4Rec with another popular Transformer-based model, namely SASRec, and show that BERT4Rec results are not consistent within these publications. To understand the reasons behind this inconsistency, we analyse the available implementations of BERT4Rec and show that we fail to reproduce results of the original BERT4Rec publication when using their default configuration parameters. However, we are able to replicate the reported results with the original code if training for a much longer amount of time (up to 30x) compared to the default configuration. We also propose our own implementation of BERT4Rec based on the Hugging Face Transformers library, which we demonstrate replicates the originally reported results on 3 out 4 datasets, while requiring up to 95 and detailed experiments, we conclude that BERT4Rec does indeed exhibit state-of-the-art effectiveness for sequential recommendation, but only when trained for a sufficient amount of time. Additionally, we show that our implementation can further benefit from adapting other Transformer architectures that are available in the Hugging Face Transformers library (e.g. using disentangled attention, as provided by DeBERTa, or larger hidden layer size cf. ALBERT).

READ FULL TEXT

page 1

page 2

page 3

page 4

research
07/26/2023

Integrating Offline Reinforcement Learning with Transformers for Sequential Recommendation

We consider the problem of sequential recommendation, where the current ...
research
04/17/2023

Attention Mixtures for Time-Aware Sequential Recommendation

Transformers emerged as powerful methods for sequential recommendation. ...
research
08/20/2023

Enhancing Transformers without Self-supervised Learning: A Loss Landscape Perspective in Sequential Recommendation

Transformer and its variants are a powerful class of architectures for s...
research
08/26/2021

The Devil is in the Detail: Simple Tricks Improve Systematic Generalization of Transformers

Recently, many datasets have been proposed to test the systematic genera...
research
12/13/2021

Sequential Recommendation with Bidirectional Chronological Augmentation of Transformer

Sequential recommendation can capture user chronological preferences fro...
research
09/17/2023

Leveraging Large Language Models for Sequential Recommendation

Sequential recommendation problems have received increasing attention in...
research
06/07/2023

GCT-TTE: Graph Convolutional Transformer for Travel Time Estimation

This paper introduces a new transformer-based model for the problem of t...

Please sign up or login with your details

Forgot password? Click here to reset