SOTAVerified

Mixture-of-Experts

Papers

Showing 691700 of 1312 papers

TitleStatusHype
LoRA-Switch: Boosting the Efficiency of Dynamic LLM Adapters via System-Algorithm Co-design0
XTrack: Multimodal Training Boosts RGB-X Video Object TrackersCode2
Yuan 2.0-M32: Mixture of Experts with Attention RouterCode2
Enhancing Fast Feed Forward Networks with Load Balancing and a Master Leaf NodeCode1
A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts0
Decomposing the Neurons: Activation Sparsity via Mixture of Experts for Continual Test Time AdaptationCode2
MoEUT: Mixture-of-Experts Universal TransformersCode2
Expert-Token Resonance: Redefining MoE Routing through Affinity-Driven Active Selection0
Revisiting MoE and Dense Speed-Accuracy Comparisons for LLM TrainingCode7
Statistical Advantages of Perturbing Cosine Router in Mixture of Experts0
Show:102550
← PrevPage 70 of 132Next →

No leaderboard results yet.