| S2MoE: Robust Sparse Mixture of Experts via Stochastic Learning | Mar 29, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Beyond Standard MoE: Mixture of Latent Experts for Resource-Efficient Language Models | Mar 29, 2025 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| Exploiting Mixture-of-Experts Redundancy Unlocks Multimodal Generative Abilities | Mar 28, 2025 | Mixture-of-ExpertsText Generation | —Unverified | 0 |
| RocketPPA: Code-Level Power, Performance, and Area Prediction via LLM and Mixture of Experts | Mar 27, 2025 | Code RepairFeature Engineering | —Unverified | 0 |
| LLaVA-CMoE: Towards Continual Mixture of Experts for Large Vision-Language Models | Mar 27, 2025 | Mixture-of-Experts | —Unverified | 0 |
| iMedImage Technical Report | Mar 27, 2025 | Anomaly DetectionDiagnostic | —Unverified | 0 |
| A multi-scale lithium-ion battery capacity prediction using mixture of experts and patch-based MLP | Mar 26, 2025 | Mixture-of-Experts | CodeCode Available | 0 |
| Reasoning Beyond Limits: Advances and Open Problems for LLMs | Mar 26, 2025 | Mixture-of-ExpertsRAG | —Unverified | 0 |
| Enhancing Multi-modal Models with Heterogeneous MoE Adapters for Fine-tuning | Mar 26, 2025 | Mixture-of-Expertsparameter-efficient fine-tuning | —Unverified | 0 |
| Optimal Scaling Laws for Efficiency Gains in a Theoretical Transformer-Augmented Sectional MoE Framework | Mar 26, 2025 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |