SOTAVerified

Mixture-of-Experts

Papers

Showing 211220 of 1312 papers

TitleStatusHype
Enhancing Fast Feed Forward Networks with Load Balancing and a Master Leaf NodeCode1
Meta-DMoE: Adapting to Domain Shift by Meta-Distillation from Mixture-of-ExpertsCode1
HyperFormer: Enhancing Entity and Relation Interaction for Hyper-Relational Knowledge Graph CompletionCode1
Contrastive Learning and Mixture of Experts Enables Precise Vector EmbeddingsCode1
MEFT: Memory-Efficient Fine-Tuning through Sparse AdapterCode1
Emergent Modularity in Pre-trained TransformersCode1
Emotion-Qwen: Training Hybrid Experts for Unified Emotion and General Vision-Language UnderstandingCode1
Jakiro: Boosting Speculative Decoding with Decoupled Multi-Head via MoECode1
MedCoT: Medical Chain of Thought via Hierarchical ExpertCode1
Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing PolicyCode1
Show:102550
← PrevPage 22 of 132Next →

No leaderboard results yet.