SOTAVerified

Mixture-of-Experts

Papers

Showing 11711180 of 1312 papers

TitleStatusHype
Not All Models Suit Expert Offloading: On Local Routing Consistency of Mixture-of-Expert ModelsCode0
Not Eliminate but Aggregate: Post-Hoc Control over Mixture-of-Experts to Address Shortcut Shifts in Natural Language UnderstandingCode0
MLP-KAN: Unifying Deep Representation and Function LearningCode0
Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-ExpertsCode0
Mixture of Nested Experts: Adaptive Processing of Visual TokensCode0
Octavius: Mitigating Task Interference in MLLMs via LoRA-MoECode0
Distribution-aware Fairness Learning in Medical Image Segmentation From A Control-Theoretic PerspectiveCode0
Mixture of Modular Experts: Distilling Knowledge from a Multilingual Teacher into Specialized Modular Language ModelsCode0
Discontinuity-Sensitive Optimal Control Learning by Mixture of ExpertsCode0
H^3Fusion: Helpful, Harmless, Honest Fusion of Aligned LLMsCode0
Show:102550
← PrevPage 118 of 132Next →

No leaderboard results yet.