| MLP Fusion: Towards Efficient Fine-tuning of Dense and Mixture-of-Experts Language Models | Jul 18, 2023 | Language ModellingMixture-of-Experts | CodeCode Available | 1 | 5 |
| PM-MOE: Mixture of Experts on Private Model Parameters for Personalized Federated Learning | Feb 1, 2025 | DenoisingFederated Learning | CodeCode Available | 1 | 5 |
| Towards Crowdsourced Training of Large Neural Networks using Decentralized Mixture-of-Experts | Feb 10, 2020 | Language ModellingMixture-of-Experts | CodeCode Available | 1 | 5 |
| ShiftAddViT: Mixture of Multiplication Primitives Towards Efficient Vision Transformer | Jun 10, 2023 | Efficient ViTsMixture-of-Experts | CodeCode Available | 1 | 5 |
| Sequence-level Semantic Representation Fusion for Recommender Systems | Feb 28, 2024 | Mixture-of-ExpertsRecommendation Systems | CodeCode Available | 1 | 5 |
| MoVEInt: Mixture of Variational Experts for Learning Human-Robot Interactions from Demonstrations | Jul 10, 2024 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| More Experts Than Galaxies: Conditionally-overlapping Experts With Biologically-Inspired Fixed Routing | Oct 10, 2024 | image-classificationImage Classification | CodeCode Available | 0 | 5 |
| MoRE-Brain: Routed Mixture of Experts for Interpretable and Generalizable Cross-Subject fMRI Visual Decoding | May 21, 2025 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| Mosaic: Data-Free Knowledge Distillation via Mixture-of-Experts for Heterogeneous Distributed Environments | May 26, 2025 | Data-free Knowledge DistillationFederated Learning | CodeCode Available | 0 | 5 |
| DeepSpeed-MoE: Advancing Mixture-of-Experts Inference and Training to Power Next-Generation AI Scale | Jan 14, 2022 | DecoderMixture-of-Experts | CodeCode Available | 0 | 5 |