| Improving Transformer Performance for French Clinical Notes Classification Using Mixture of Experts on a Limited Dataset | Mar 22, 2023 | Mixture-of-Expertstext-classification | —Unverified | 0 | 0 |
| Astrea: A MOE-based Visual Understanding Model with Progressive Alignment | Mar 12, 2025 | Contrastive LearningCross-Modal Retrieval | —Unverified | 0 | 0 |
| A Survey on Dynamic Neural Networks for Natural Language Processing | Feb 15, 2022 | Dynamic neural networksMixture-of-Experts | —Unverified | 0 | 0 |
| A Survey on Model MoErging: Recycling and Routing Among Specialized Experts for Collaborative Learning | Aug 13, 2024 | Mixture-of-ExpertsSurvey | —Unverified | 0 | 0 |
| A Theoretical View on Sparsely Activated Networks | Aug 8, 2022 | Mixture-of-Experts | —Unverified | 0 | 0 |
| AT-MoE: Adaptive Task-planning Mixture of Experts via LoRA Approach | Oct 12, 2024 | Mixture-of-ExpertsTask Planning | —Unverified | 0 | 0 |
| A Tree Architecture of LSTM Networks for Sequential Regression with Missing Data | May 22, 2020 | Mixture-of-Expertsregression | —Unverified | 0 | 0 |
| Attention Weighted Mixture of Experts with Contrastive Learning for Personalized Ranking in E-commerce | Jun 8, 2023 | Contrastive LearningMixture-of-Experts | —Unverified | 0 | 0 |
| A Two-Phase Deep Learning Framework for Adaptive Time-Stepping in High-Speed Flow Modeling | Jun 9, 2025 | Mixture-of-Experts | —Unverified | 0 | 0 |
| A Unified Approach to Universal Prediction: Generalized Upper and Lower Bounds | Nov 25, 2013 | Learning TheoryMixture-of-Experts | —Unverified | 0 | 0 |