SOTAVerified

Mixture-of-Experts

Papers

Showing 901910 of 1312 papers

TitleStatusHype
Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing PolicyCode1
MoCaE: Mixture of Calibrated Experts Significantly Improves Object DetectionCode1
LLMCarbon: Modeling the end-to-end Carbon Footprint of Large Language ModelsCode1
Statistical Perspective of Top-K Sparse Softmax Gating Mixture of Experts0
Pushing Mixture of Experts to the Limit: Extremely Parameter Efficient MoE for Instruction TuningCode2
Mobile V-MoEs: Scaling Down Vision Transformers via Sparse Mixture-of-Experts0
Exploring Sparse MoE in GANs for Text-conditioned Image SynthesisCode1
Learning multi-modal generative models with permutation-invariant encoders and tighter variational objectivesCode0
Task-Based MoE for Multitask Multilingual Machine Translation0
SwapMoE: Serving Off-the-shelf MoE-based Large Language Models with Tunable Memory Budget0
Show:102550
← PrevPage 91 of 132Next →

No leaderboard results yet.