| Terminating Differentiable Tree Experts | Jul 2, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Efficient Expert Pruning for Sparse Mixture-of-Experts Language Models: Enhancing Performance and Reducing Inference Costs | Jul 1, 2024 | GPUMixture-of-Experts | CodeCode Available | 1 |
| Investigating the potential of Sparse Mixtures-of-Experts for multi-domain neural machine translation | Jul 1, 2024 | Machine TranslationMixture-of-Experts | —Unverified | 0 |
| Sparse Diffusion Policy: A Sparse, Reusable, and Flexible Policy for Robot Learning | Jul 1, 2024 | Continual LearningMixture-of-Experts | —Unverified | 0 |
| Solving Token Gradient Conflict in Mixture-of-Experts for Large Vision-Language Model | Jun 28, 2024 | Language ModelingLanguage Modelling | CodeCode Available | 1 |
| LEMoE: Advanced Mixture of Experts Adaptor for Lifelong Model Editing of Large Language Models | Jun 28, 2024 | Mixture-of-ExpertsModel Editing | —Unverified | 0 |
| A Teacher Is Worth A Million Instructions | Jun 27, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| Towards Personalized Federated Multi-Scenario Multi-Task Recommendation | Jun 27, 2024 | Federated LearningMixture-of-Experts | —Unverified | 0 |
| A Survey on Mixture of Experts | Jun 26, 2024 | In-Context LearningMixture-of-Experts | CodeCode Available | 3 |
| SC-MoE: Switch Conformer Mixture of Experts for Unified Streaming and Non-streaming Code-Switching ASR | Jun 26, 2024 | Automatic Speech RecognitionAutomatic Speech Recognition (ASR) | —Unverified | 0 |