SOTAVerified

Mixture-of-Experts

Papers

Showing 476500 of 1312 papers

TitleStatusHype
ViMoE: An Empirical Study of Designing Vision Mixture-of-Experts0
LoRA-IR: Taming Low-Rank Experts for Efficient All-in-One Image RestorationCode2
MENTOR: Mixture-of-Experts Network with Task-Oriented Perturbation for Visual Reinforcement Learning0
MomentumSMoE: Integrating Momentum into Sparse Mixture of ExpertsCode1
ST-MoE-BERT: A Spatial-Temporal Mixture-of-Experts Framework for Long-Term Cross-City Mobility PredictionCode1
Enhancing Generalization in Sparse Mixture of Experts Models: The Case for Increased Expert Activation in Compositional Tasks0
On the Risk of Evidence Pollution for Malicious Social Text Detection in the Era of LLMs0
EPS-MoE: Expert Pipeline Scheduler for Cost-Efficient MoE Inference0
Understanding Expert Structures on Minimax Parameter Estimation in Contaminated Mixture of Experts0
Transformer Layer Injection: A Novel Approach for Efficient Upscaling of Large Language Models0
MoE-Pruner: Pruning Mixture-of-Experts Large Language Model using the Hints from Its Router0
MoH: Multi-Head Attention as Mixture-of-Head AttentionCode4
Quadratic Gating Functions in Mixture of Experts: A Statistical Insight0
GaVaMoE: Gaussian-Variational Gated Mixture of Experts for Explainable RecommendationCode1
AlphaLoRA: Assigning LoRA Experts Based on Layer Training QualityCode1
Ada-K Routing: Boosting the Efficiency of MoE-based LLMs0
Efficiently Democratizing Medical LLMs for 50 Languages via a Mixture of Language Family ExpertsCode2
Learning to Ground VLMs without Forgetting0
Your Mixture-of-Experts LLM Is Secretly an Embedding Model For FreeCode2
Scalable Multi-Domain Adaptation of Language Models using Modular Experts0
Mixture of Experts Made Personalized: Federated Prompt Learning for Vision-Language ModelsCode1
Moirai-MoE: Empowering Time Series Foundation Models with Sparse Mixture of ExpertsCode5
ContextWIN: Whittle Index Based Mixture-of-Experts Neural Model For Restless Bandits Via Deep RL0
MoIN: Mixture of Introvert Experts to Upcycle an LLM0
AT-MoE: Adaptive Task-planning Mixture of Experts via LoRA Approach0
Show:102550
← PrevPage 20 of 53Next →

No leaderboard results yet.