| Approximation Rates and VC-Dimension Bounds for (P)ReLU MLP Mixture of Experts | Feb 5, 2024 | GPUMixture-of-Experts | —Unverified | 0 |
| On Least Square Estimation in Softmax Gating Mixture of Experts | Feb 5, 2024 | Mixture-of-Experts | —Unverified | 0 |
| FuseMoE: Mixture-of-Experts Transformers for Fleximodal Fusion | Feb 5, 2024 | Missing ElementsMixture-of-Experts | —Unverified | 0 |
| CompeteSMoE - Effective Training of Sparse Mixture of Experts via Competition | Feb 4, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| pFedMoE: Data-Level Personalization with Mixture of Experts for Model-Heterogeneous Personalized Federated Learning | Feb 2, 2024 | Federated LearningMixture-of-Experts | CodeCode Available | 0 |
| MoDE: A Mixture-of-Experts Model with Mutual Distillation among the Experts | Jan 31, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Explainable data-driven modeling via mixture of experts: towards effective blending of grey and black-box models | Jan 30, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Checkmating One, by Using Many: Combining Mixture of Experts with MCTS to Improve in Chess | Jan 30, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| LLaVA-MoLE: Sparse Mixture of LoRA Experts for Mitigating Data Conflicts in Instruction Finetuning MLLMs | Jan 29, 2024 | Language ModellingLarge Language Model | —Unverified | 0 |
| Routers in Vision Mixture of Experts: An Empirical Study | Jan 29, 2024 | Language ModelingLanguage Modelling | —Unverified | 0 |