SOTAVerified

Mixture-of-Experts

Papers

Showing 131140 of 1312 papers

TitleStatusHype
Outrageously Large Neural Networks: The Sparsely-Gated Mixture-of-Experts LayerCode2
Learning to Skip the Middle Layers of TransformersCode1
Structural Similarity-Inspired Unfolding for Lightweight Image Super-ResolutionCode1
SPACE: Your Genomic Profile Predictor is a Powerful DNA Foundation ModelCode1
Mastering Massive Multi-Task Reinforcement Learning via Mixture-of-Expert Decision TransformerCode1
FLAME-MoE: A Transparent End-to-End Research Platform for Mixture-of-Experts Language ModelsCode1
ThanoRA: Task Heterogeneity-Aware Multi-Task Low-Rank AdaptationCode1
JanusDNA: A Powerful Bi-directional Hybrid DNA Foundation ModelCode1
U-SAM: An audio language Model for Unified Speech, Audio, and Music UnderstandingCode1
Occult: Optimizing Collaborative Communication across Experts for Accelerated Parallel MoE Training and InferenceCode1
Show:102550
← PrevPage 14 of 132Next →

No leaderboard results yet.