| Task-Specific Expert Pruning for Sparse Mixture-of-Experts | Jun 1, 2022 | Mixture-of-Experts | —Unverified | 0 | 0 |
| Team Deep Mixture of Experts for Distributed Power Control | Jul 28, 2020 | Mixture-of-Expertsspeech-recognition | —Unverified | 0 | 0 |
| Terminating Differentiable Tree Experts | Jul 2, 2024 | Mixture-of-Experts | —Unverified | 0 | 0 |
| The Empirical Impact of Reducing Symmetries on the Performance of Deep Ensembles and MoE | Feb 24, 2025 | Linear Mode ConnectivityMixture-of-Experts | —Unverified | 0 | 0 |
| The Labyrinth of Links: Navigating the Associative Maze of Multi-modal LLMs | Oct 2, 2024 | BenchmarkingHallucination | —Unverified | 0 | 0 |
| Theory of Mixture-of-Experts for Mobile Edge Computing | Dec 20, 2024 | Computational EfficiencyContinual Learning | —Unverified | 0 | 0 |
| Theory on Mixture-of-Experts in Continual Learning | Jun 24, 2024 | Continual LearningMixture-of-Experts | —Unverified | 0 | 0 |
| The power of fine-grained experts: Granularity boosts expressivity in Mixture of Experts | May 11, 2025 | Mixture-of-Experts | —Unverified | 0 | 0 |
| The Ultimate Guide to Fine-Tuning LLMs from Basics to Breakthroughs: An Exhaustive Review of Technologies, Research, Best Practices, Applied Research Challenges and Opportunities | Aug 23, 2024 | Computational EfficiencyInference Optimization | —Unverified | 0 | 0 |
| THOR-MoE: Hierarchical Task-Guided and Context-Responsive Routing for Neural Machine Translation | May 20, 2025 | Machine TranslationMixture-of-Experts | —Unverified | 0 | 0 |