When Newer is Not Better: Does Deep Learning Really Benefit Recommendation From Implicit Feedback?

05/02/2023
by   Yushun Dong, et al.
0

In recent years, neural models have been repeatedly touted to exhibit state-of-the-art performance in recommendation. Nevertheless, multiple recent studies have revealed that the reported state-of-the-art results of many neural recommendation models cannot be reliably replicated. A primary reason is that existing evaluations are performed under various inconsistent protocols. Correspondingly, these replicability issues make it difficult to understand how much benefit we can actually gain from these neural models. It then becomes clear that a fair and comprehensive performance comparison between traditional and neural models is needed. Motivated by these issues, we perform a large-scale, systematic study to compare recent neural recommendation models against traditional ones in top-n recommendation from implicit data. We propose a set of evaluation strategies for measuring memorization performance, generalization performance, and subgroup-specific performance of recommendation models. We conduct extensive experiments with 13 popular recommendation models (including two neural models and 11 traditional ones as baselines) on nine commonly used datasets. Our experiments demonstrate that even with extensive hyper-parameter searches, neural models do not dominate traditional models in all aspects, e.g., they fare worse in terms of average HitRate. We further find that there are areas where neural models seem to outperform non-neural models, for example, in recommendation diversity and robustness between different subgroups of users and items. Our work illuminates the relative advantages and disadvantages of neural models in recommendation and is therefore an important step towards building better recommender systems.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
03/02/2022

Top-N Recommendation Algorithms: A Quest for the State-of-the-Art

Research on recommender systems algorithms, like other areas of applied ...
research
03/15/2022

Non-neural Models Matter: A Re-evaluation of Neural Referring Expression Generation Systems

In recent years, neural models have often outperformed rule-based and cl...
research
08/29/2022

Understanding Diversity in Session-Based Recommendation

Current session-based recommender systems (SBRSs) mainly focus on maximi...
research
06/11/2018

Distributed Evaluations: Ending Neural Point Metrics

With the rise of neural models across the field of information retrieval...
research
04/02/2021

Humor@IITK at SemEval-2021 Task 7: Large Language Models for Quantifying Humor and Offensiveness

Humor and Offense are highly subjective due to multiple word senses, cul...
research
04/07/2018

Evaluating historical text normalization systems: How well do they generalize?

We highlight several issues in the evaluation of historical text normali...
research
11/03/2021

Rethinking the Image Feature Biases Exhibited by Deep CNN Models

In recent years, convolutional neural networks (CNNs) have been applied ...

Please sign up or login with your details

Forgot password? Click here to reset