| Sparse Mixers: Combining MoE and Mixing to build a more efficient BERT | May 24, 2022 | Mixture-of-Experts | —Unverified | 0 | 0 |
| Sparse Mixture of Experts as Unified Competitive Learning | Mar 29, 2025 | Language ModelingLanguage Modelling | —Unverified | 0 | 0 |
| Sparse Mixture-of-Experts for Non-Uniform Noise Reduction in MRI Images | Jan 24, 2025 | DenoisingDiagnostic | —Unverified | 0 | 0 |
| Cross-token Modeling with Conditional Computation | Sep 5, 2021 | Computational EfficiencyImage Classification | —Unverified | 0 | 0 |
| Sparse Upcycling: Inference Inefficient Finetuning | Nov 13, 2024 | Mixture-of-Experts | —Unverified | 0 | 0 |
| Sparse Video Representation Using Steered Mixture-of-Experts With Global Motion Compensation | Sep 13, 2022 | Mixture-of-ExpertsMotion Compensation | —Unverified | 0 | 0 |
| Sparsity-Constrained Optimal Transport | Sep 30, 2022 | Mixture-of-Experts | —Unverified | 0 | 0 |
| Speculative MoE: Communication Efficient Parallel MoE Inference with Speculative Token and Expert Pre-scheduling | Mar 6, 2025 | Mixture-of-ExpertsScheduling | —Unverified | 0 | 0 |
| SpeechMatrix: A Large-Scale Mined Corpus of Multilingual Speech-to-Speech Translations | Nov 8, 2022 | Mixture-of-ExpertsSpeech-to-Speech Translation | —Unverified | 0 | 0 |
| SpeechMoE2: Mixture-of-Experts Model with Improved Routing | Nov 23, 2021 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 | 0 |