Enhancing Dyadic Relations with Homogeneous Graphs for Multimodal Recommendation

01/28/2023
by   Hongyu Zhou, et al.
0

User interaction data in recommender systems is a form of dyadic relation that reflects the preferences of users with items. Learning the representations of these two discrete sets of objects, users and items, is critical for recommendation. Recent multimodal recommendation models leveraging multimodal features (e.g., images and text descriptions) have been demonstrated to be effective in improving recommendation accuracy. However, state-of-the-art models enhance the dyadic relations between users and items by considering either user-user or item-item relations, leaving the high-order relations of the other side (i.e., users or items) unexplored. Furthermore, we experimentally reveal that the current multimodality fusion methods in the state-of-the-art models may degrade their recommendation performance. That is, without tainting the model architectures, these models can achieve even better recommendation accuracy with uni-modal information. On top of the finding, we propose a model that enhances the dyadic relations by learning Dual RepresentAtions of both users and items via constructing homogeneous Graphs for multimOdal recommeNdation. We name our model as DRAGON. Specifically, DRAGON constructs the user-user graph based on the commonly interacted items and the item-item graph from item multimodal features. It then utilizes graph learning on both the user-item heterogeneous graph and the homogeneous graphs (user-user and item-item) to obtain the dual representations of users and items. To capture information from each modality, DRAGON employs a simple yet effective fusion method, attentive concatenation, to derive the representations of users and items. Extensive experiments on three public datasets and seven baselines show that DRAGON can outperform the strongest baseline by 22.03 Various ablation studies are conducted on DRAGON to validate its effectiveness.

READ FULL TEXT
research
04/19/2021

Mining Latent Structures for Multimedia Recommendation

Multimedia content is of predominance in the modern Web era. Investigati...
research
04/24/2023

Attention-guided Multi-step Fusion: A Hierarchical Fusion Network for Multimodal Recommendation

The main idea of multimodal recommendation is the rational utilization o...
research
01/24/2022

Dual Preference Distribution Learning for Item Recommendation

Recommender systems can automatically recommend users items that they pr...
research
07/12/2018

Multi-Perspective Neural Architecture for Recommendation System

Currently, there starts a research trend to leverage neural architecture...
research
06/09/2020

I know why you like this movie: Interpretable Efficient Multimodal Recommender

Recently, the Efficient Manifold Density Estimator (EMDE) model has been...
research
11/13/2022

A Tale of Two Graphs: Freezing and Denoising Graph Structures for Multimodal Recommendation

Multimodal recommender systems utilizing multimodal features (e.g. image...
research
07/13/2022

Bootstrap Latent Representations for Multi-modal Recommendation

This paper studies the multi-modal recommendation problem, where the ite...

Please sign up or login with your details

Forgot password? Click here to reset