SOTAVerified

Mixture-of-Experts

Papers

Showing 151175 of 1312 papers

TitleStatusHype
StableFusion: Continual Video Retrieval via Frame AdaptationCode1
Samoyeds: Accelerating MoE Models with Structured Sparsity Leveraging Sparse Tensor CoresCode1
Question-Aware Gaussian Experts for Audio-Visual Question AnsweringCode1
Small but Mighty: Enhancing Time Series Forecasting with Lightweight LLMsCode1
MX-Font++: Mixture of Heterogeneous Aggregation Experts for Few-shot Font GenerationCode1
R2-T2: Re-Routing in Test-Time for Multimodal Mixture-of-ExpertsCode1
ChatVLA: Unified Multimodal Understanding and Robot Control with Vision-Language-Action ModelCode1
Heterogeneous Mixture of Experts for Remote Sensing Image Super-ResolutionCode1
Jakiro: Boosting Speculative Decoding with Decoupled Multi-Head via MoECode1
CMoE: Fast Carving of Mixture-of-Experts for Efficient LLM InferenceCode1
UniGraph2: Learning a Unified Embedding Space to Bind Multimodal GraphsCode1
PM-MOE: Mixture of Experts on Private Model Parameters for Personalized Federated LearningCode1
FreqMoE: Enhancing Time Series Forecasting through Frequency Decomposition Mixture of ExpertsCode1
Hierarchical Time-Aware Mixture of Experts for Multi-Modal Sequential RecommendationCode1
Modality Interactive Mixture-of-Experts for Fake News DetectionCode1
MoGERNN: An Inductive Traffic Predictor for Unobserved Locations in Dynamic Sensing NetworksCode1
Transforming Vision Transformer: Towards Efficient Multi-Task Asynchronous LearningCode1
BrainMAP: Learning Multiple Activation Pathways in Brain NetworksCode1
MedCoT: Medical Chain of Thought via Hierarchical ExpertCode1
Wonderful Matrices: Combining for a More Efficient and Effective Foundation Model ArchitectureCode1
RSUniVLM: A Unified Vision Language Model for Remote Sensing via Granularity-oriented Mixture of ExpertsCode1
SAME: Learning Generic Language-Guided Visual Navigation with State-Adaptive Mixture of ExpertsCode1
Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer PruningCode1
Awaker2.5-VL: Stably Scaling MLLMs with Parameter-Efficient Mixture of ExpertsCode1
LIBMoE: A Library for comprehensive benchmarking Mixture of Experts in Large Language ModelsCode1
Show:102550
← PrevPage 7 of 53Next →

No leaderboard results yet.