| Connector-S: A Survey of Connectors in Multi-modal Large Language Models | Feb 17, 2025 | Mixture-of-ExpertsSurvey | —Unverified | 0 |
| Configurable Foundation Models: Building LLMs from a Modular Perspective | Sep 4, 2024 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| 3D-MoE: A Mixture-of-Experts Multi-modal LLM for 3D Vision and Pose Diffusion via Rectified Flow | Jan 28, 2025 | Instruction FollowingMixture-of-Experts | —Unverified | 0 |
| Conditional computation in neural networks: principles and research trends | Mar 12, 2024 | Mixture-of-Expertsscientific discovery | —Unverified | 0 |
| On DeepSeekMoE: Statistical Benefits of Shared Experts and Normalized Sigmoid Gating | May 16, 2025 | Language ModelingLanguage Modelling | —Unverified | 0 |
| On the Adaptation to Concept Drift for CTR Prediction | Apr 1, 2022 | Click-Through Rate PredictionIncremental Learning | —Unverified | 0 |
| A Review of Sparse Expert Models in Deep Learning | Sep 4, 2022 | Deep LearningMixture-of-Experts | —Unverified | 0 |
| Complexity Experts are Task-Discriminative Learners for Any Image Restoration | Nov 27, 2024 | AttributeBlind All-in-One Image Restoration | —Unverified | 0 |
| Filtered not Mixed: Stochastic Filtering-Based Online Gating for Mixture of Large Language Models | Jun 5, 2024 | Mixture-of-ExpertsTime Series | —Unverified | 0 |
| A Review of DeepSeek Models' Key Innovative Techniques | Mar 14, 2025 | Mixture-of-Expertsreinforcement-learning | —Unverified | 0 |