On the Effectiveness of Sampled Softmax Loss for Item Recommendation

01/07/2022
by   Jiancan Wu, et al.
0

Learning objectives of recommender models remain largely unexplored. Most methods routinely adopt either pointwise or pairwise loss to train the model parameters, while rarely pay attention to softmax loss due to the high computational cost. Sampled softmax loss emerges as an efficient substitute for softmax loss. Its special case, InfoNCE loss, has been widely used in self-supervised learning and exhibited remarkable performance for contrastive learning. Nonetheless, limited studies use sampled softmax loss as the learning objective to train the recommender. Worse still, none of them explore its properties and answer "Does sampled softmax loss suit for item recommendation?" and "What are the conceptual advantages of sampled softmax loss, as compared with the prevalent losses?", to the best of our knowledge. In this work, we aim to better understand sampled softmax loss for item recommendation. Specifically, we first theoretically reveal three model-agnostic advantages: (1) mitigating popularity bias, which is beneficial to long-tail recommendation; (2) mining hard negative samples, which offers informative gradients to optimize model parameters; and (3) maximizing the ranking metric, which facilitates top-K performance. Moreover, we probe the model-specific characteristics on the top of various recommenders. Experimental results suggest that sampled softmax loss is more friendly to history and graph-based recommenders (e.g., SVD++ and LightGCN), but performs poorly for ID-based models (e.g., MF). We ascribe this to its shortcoming in learning representation magnitude, making the combination with the models that are also incapable of adjusting representation magnitude learn poor representations. In contrast, the history- and graph-based models, which naturally adjust representation magnitude according to node degree, are able to compensate for the shortcoming of sampled softmax loss.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
04/10/2020

Efficient Sampled Softmax for Tensorflow

This short paper discusses an efficient implementation of sampled softma...
research
10/21/2020

Self-supervised Graph Learning for Recommendation

Representation learning on user-item graph for recommendation has evolve...
research
07/07/2021

SelfCF: A Simple Framework for Self-supervised Collaborative Filtering

Collaborative filtering (CF) is widely used to learn an informative late...
research
03/09/2022

Efficient Image Representation Learning with Federated Sampled Softmax

Learning image representations on decentralized data can bring many bene...
research
05/30/2022

Cache-Augmented Inbatch Importance Resampling for Training Recommender Retriever

Recommender retrievers aim to rapidly retrieve a fraction of items from ...
research
01/05/2021

Contrastive Learning for Recommender System

Recommender systems, which analyze users' preference patterns to suggest...
research
10/11/2021

Breaking the Softmax Bottleneck for Sequential Recommender Systems with Dropout and Decoupling

The Softmax bottleneck was first identified in language modeling as a th...

Please sign up or login with your details

Forgot password? Click here to reset