SOTAVerified

Multimodal Recommendation

The multimodal recommendation task involves developing systems that leverage and integrate multiple types of data—such as text, images, audio, and user interactions—to predict and suggest items that align with a user's preferences. Unlike traditional recommendation approaches that rely on a single data modality, multimodal recommendation harnesses the diverse information from various sources to create richer and more nuanced representations of both users and items. This integration enables the system to understand and capture complex relationships and attributes across different data types, thereby enhancing the accuracy and relevance of the recommendations. The primary goal is to provide personalized suggestions by effectively merging and processing heterogeneous data to better match users with items they are likely to engage with or find valuable.

Papers

Showing 110 of 59 papers

TitleStatusHype
MDVT: Enhancing Multimodal Recommendation with Model-Agnostic Multimodal-Driven Virtual Triplets0
A Survey on Large Language Models in Multimodal Recommender Systems0
Quadratic Interest Network for Multimodal Click-Through Rate PredictionCode1
Modality Reliability Guided Multimodal Recommendation0
HistLLM: A Unified Framework for LLM-Based Multimodal Recommendation with User History Encoding and Compression0
Multi-Modal Hypergraph Enhanced LLM Learning for Recommendation0
COHESION: Composite Graph Convolutional Network with Dual-Stage Fusion for Multimodal RecommendationCode1
Training-Free Graph Filtering via Multimodal Feature Refinement for Extremely Fast Multimodal Recommendation0
Bridging Domain Gaps between Pretrained Multimodal Models and Recommendations0
Collaborative Filtering Meets Spectrum Shift: Connecting User-Item Interaction with Graph-Structured Side InformationCode0
Show:102550
← PrevPage 1 of 6Next →

No leaderboard results yet.