SOTAVerified

Mixture-of-Experts

Papers

Showing 101110 of 1312 papers

TitleStatusHype
FuxiMT: Sparsifying Large Language Models for Chinese-Centric Multilingual Machine Translation0
StPR: Spatiotemporal Preservation and Routing for Exemplar-Free Video Class-Incremental Learning0
Two Experts Are All You Need for Steering Thinking: Reinforcing Cognitive Effort in MoE Reasoning Models Without Additional Training0
U-SAM: An audio language Model for Unified Speech, Audio, and Music UnderstandingCode1
Scaling and Enhancing LLM-based AVSR: A Sparse Mixture of Projectors Approach0
Occult: Optimizing Collaborative Communication across Experts for Accelerated Parallel MoE Training and InferenceCode1
Model Selection for Gaussian-gated Gaussian Mixture of Experts Using Dendrograms of Mixing Measures0
True Zero-Shot Inference of Dynamical Systems Preserving Long-Term Statistics0
CompeteSMoE -- Statistically Guaranteed Mixture of Experts Training via CompetitionCode0
Seeing the Unseen: How EMoE Unveils Bias in Text-to-Image Diffusion Models0
Show:102550
← PrevPage 11 of 132Next →

No leaderboard results yet.