Do LLMs Understand User Preferences? Evaluating LLMs On User Rating Prediction

05/10/2023
by   Wang-Cheng Kang, et al.
0

Large Language Models (LLMs) have demonstrated exceptional capabilities in generalizing to new tasks in a zero-shot or few-shot manner. However, the extent to which LLMs can comprehend user preferences based on their previous behavior remains an emerging and still unclear research question. Traditionally, Collaborative Filtering (CF) has been the most effective method for these tasks, predominantly relying on the extensive volume of rating data. In contrast, LLMs typically demand considerably less data while maintaining an exhaustive world knowledge about each item, such as movies or products. In this paper, we conduct a thorough examination of both CF and LLMs within the classic task of user rating prediction, which involves predicting a user's rating for a candidate item based on their past ratings. We investigate various LLMs in different sizes, ranging from 250M to 540B parameters and evaluate their performance in zero-shot, few-shot, and fine-tuning scenarios. We conduct comprehensive analysis to compare between LLMs and strong CF methods, and find that zero-shot LLMs lag behind traditional recommender models that have the access to user interaction data, indicating the importance of user interaction data. However, through fine-tuning, LLMs achieve comparable or even better performance with only a small fraction of the training data, demonstrating their potential through data efficiency.

READ FULL TEXT

page 1

page 2

page 3

page 4

research
09/03/2023

Pre-trained Neural Recommenders: A Transferable Zero-Shot Framework for Recommendation Systems

Modern neural collaborative filtering techniques are critical to the suc...
research
06/09/2023

How Does Fine-Tuning Impact Out-of-Distribution Detection for Vision-Language Models?

Recent large vision-language models such as CLIP have shown remarkable o...
research
07/26/2023

Large Language Models are Competitive Near Cold-start Recommenders for Language- and Item-based Preferences

Traditional recommender systems leverage users' item preference history ...
research
12/21/2022

From Images to Textual Prompts: Zero-shot VQA with Frozen Large Language Models

Large language models (LLMs) have demonstrated excellent zero-shot gener...
research
09/13/2023

Scaled Prompt-Tuning for Few-Shot Natural Language Generation

The increasingly Large Language Models (LLMs) demonstrate stronger langu...
research
02/01/2023

KNNs of Semantic Encodings for Rating Prediction

This paper explores a novel application of textual semantic similarity t...
research
04/05/2022

"Does it come in black?" CLIP-like models are zero-shot recommenders

Product discovery is a crucial component for online shopping. However, i...

Please sign up or login with your details

Forgot password? Click here to reset