SOTAVerified

Multimodal Recommendation

The multimodal recommendation task involves developing systems that leverage and integrate multiple types of data—such as text, images, audio, and user interactions—to predict and suggest items that align with a user's preferences. Unlike traditional recommendation approaches that rely on a single data modality, multimodal recommendation harnesses the diverse information from various sources to create richer and more nuanced representations of both users and items. This integration enables the system to understand and capture complex relationships and attributes across different data types, thereby enhancing the accuracy and relevance of the recommendations. The primary goal is to provide personalized suggestions by effectively merging and processing heterogeneous data to better match users with items they are likely to engage with or find valuable.

Papers

Showing 150 of 59 papers

TitleStatusHype
Modality-Independent Graph Neural Networks with Global Transformers for Multimodal RecommendationCode2
A Comprehensive Survey on Multimodal Recommender Systems: Taxonomy, Evaluation, and Future DirectionsCode2
Mining Latent Structures for Multimedia RecommendationCode1
Mirror Gradient: Towards Robust Multimodal Recommender Systems via Exploring Flat Local MinimaCode1
A Tale of Two Graphs: Freezing and Denoising Graph Structures for Multimodal RecommendationCode1
Modality-Balanced Learning for Multimedia RecommendationCode1
LightGT: A Light Graph Transformer for Multimedia RecommendationCode1
MENTOR: Multi-level Self-supervised Learning for Multimodal RecommendationCode1
Spectrum-based Modality Representation Fusion Graph Convolutional Network for Multimodal RecommendationCode1
Generating with Fairness: A Modality-Diffused Counterfactual Framework for Incomplete Multimodal RecommendationsCode1
GUME: Graphs and User Modalities Enhancement for Long-Tail Multimodal RecommendationCode1
Train Once, Deploy Anywhere: Matryoshka Representation Learning for Multimodal RecommendationCode1
Harnessing Multimodal Large Language Models for Multimodal Sequential RecommendationCode1
AlignRec: Aligning and Training in Multimodal RecommendationsCode1
COHESION: Composite Graph Convolutional Network with Dual-Stage Fusion for Multimodal RecommendationCode1
LGMRec: Local and Global Graph Learning for Multimodal RecommendationCode1
Disentangled Graph Variational Auto-Encoder for Multimodal Recommendation with InterpretabilityCode1
Beyond Graph Convolution: Multimodal Recommendation with Topology-aware MLPsCode1
Ducho 2.0: Towards a More Up-to-Date Unified Framework for the Extraction of Multimodal Features in RecommendationCode1
Ducho: A Unified Framework for the Extraction of Multimodal Features in RecommendationCode1
End-to-end training of Multimodal Model and ranking ModelCode1
Enhancing Dyadic Relations with Homogeneous Graphs for Multimodal RecommendationCode1
Quadratic Interest Network for Multimodal Click-Through Rate PredictionCode1
Causality-Inspired Fair Representation Learning for Multimodal RecommendationCode1
MMRec: Simplifying Multimodal RecommendationCode0
Do We Really Need to Drop Items with Missing Modalities in Multimodal Recommendation?Code0
MMGCN: Multi-modal Graph Convolution Network for Personalized Recommendation of Micro-videoCode0
A Multimodal Single-Branch Embedding Network for Recommendation in Cold-Start and Missing Modality ScenariosCode0
Semantic-Guided Feature Distillation for Multimodal RecommendationCode0
Ducho meets Elliot: Large-scale Benchmarks for Multimodal RecommendationCode0
Dynamic Multimodal Fusion via Meta-Learning Towards Micro-Video RecommendationCode0
Collaborative Filtering Meets Spectrum Shift: Connecting User-Item Interaction with Graph-Structured Side InformationCode0
STAIR: Manipulating Collaborative and Multimodal Information for E-Commerce RecommendationCode0
X-Reflect: Cross-Reflection Prompting for Multimodal Recommendation0
A Survey on Large Language Models in Multimodal Recommender Systems0
ATFLRec: A Multimodal Recommender System with Audio-Text Fusion and Low-Rank Adaptation via Instruction-Tuned Large Language Model0
Attention-guided Multi-step Fusion: A Hierarchical Fusion Network for Multimodal Recommendation0
Attribute-driven Disentangled Representation Learning for Multimodal Recommendation0
Bridging Domain Gaps between Pretrained Multimodal Models and Recommendations0
Dealing with Missing Modalities in Multimodal Recommendation: a Feature Propagation-based Approach0
Don't Lose Yourself: Boosting Multimodal Recommendation via Reducing Node-neighbor Discrepancy in Graph Convolutional Network0
DREAM: A Dual Representation Learning Model for Multimodal Recommendation0
HistLLM: A Unified Framework for LLM-Based Multimodal Recommendation with User History Encoding and Compression0
ID Embedding as Subtle Features of Content and Structure for Multimodal Recommendation0
Knowledge Soft Integration for Multimodal Recommendation0
Learning ID-free Item Representation with Token Crossing for Multimodal Recommendation0
MDVT: Enhancing Multimodal Recommendation with Model-Agnostic Multimodal-Driven Virtual Triplets0
MMGRec: Multimodal Generative Recommendation with Transformer Model0
Modality Reliability Guided Multimodal Recommendation0
Multimodal Graph Neural Network for Recommendation with Dynamic De-redundancy and Modality-Guided Feature De-noisy0
Show:102550
← PrevPage 1 of 2Next →

No leaderboard results yet.