SOTAVerified

Multimodal Recommendation

The multimodal recommendation task involves developing systems that leverage and integrate multiple types of data—such as text, images, audio, and user interactions—to predict and suggest items that align with a user's preferences. Unlike traditional recommendation approaches that rely on a single data modality, multimodal recommendation harnesses the diverse information from various sources to create richer and more nuanced representations of both users and items. This integration enables the system to understand and capture complex relationships and attributes across different data types, thereby enhancing the accuracy and relevance of the recommendations. The primary goal is to provide personalized suggestions by effectively merging and processing heterogeneous data to better match users with items they are likely to engage with or find valuable.

Papers

Showing 5159 of 59 papers

TitleStatusHype
Rec-GPT4V: Multimodal Recommendation with Large Vision-Language Models0
Attribute-driven Disentangled Representation Learning for Multimodal Recommendation0
ID Embedding as Subtle Features of Content and Structure for Multimodal Recommendation0
Semantic-Guided Feature Distillation for Multimodal RecommendationCode0
Multimodal Recommendation Dialog with Subjective Preference: A New Challenge and Benchmark0
Knowledge Soft Integration for Multimodal Recommendation0
Attention-guided Multi-step Fusion: A Hierarchical Fusion Network for Multimodal Recommendation0
MMRec: Simplifying Multimodal RecommendationCode0
MMGCN: Multi-modal Graph Convolution Network for Personalized Recommendation of Micro-videoCode0
Show:102550
← PrevPage 3 of 3Next →

No leaderboard results yet.