SOTAVerified

Mixture-of-Experts

Papers

Showing 9911000 of 1312 papers

TitleStatusHype
Memory-efficient NLLB-200: Language-specific Expert Pruning of a Massively Multilingual Machine Translation Model0
Fixing MoE Over-Fitting on Low-Resource Languages in Multilingual Machine Translation0
Mod-Squad: Designing Mixture of Experts As Modular Multi-Task Learners0
SMILE: Scaling Mixture-of-Experts with Efficient Bi-level Routing0
Sparse Upcycling: Training Mixture-of-Experts from Dense CheckpointsCode2
Incorporating Polar Field Data for Improved Solar Flare Prediction0
Named Entity and Relation Extraction with Multi-Modal Retrieval0
MegaBlocks: Efficient Sparse Training with Mixture-of-ExpertsCode3
Automatically Extracting Information in Medical Dialogue: Expert System And Attention for Labelling0
Mixture of Decision Trees for Interpretable Machine LearningCode1
Show:102550
← PrevPage 100 of 132Next →

No leaderboard results yet.