| MoTE: Mixture of Ternary Experts for Memory-efficient Large Multimodal Models | Jun 17, 2025 | Mixture-of-ExpertsQuantization | —Unverified | 0 |
| NeuroMoE: A Transformer-Based Mixture-of-Experts Framework for Multi-Modal Neurological Disorder Classification | Jun 17, 2025 | DiagnosticMixture-of-Experts | —Unverified | 0 |
| Ring-lite: Scalable Reasoning via C3PO-Stabilized Reinforcement Learning for LLMs | Jun 17, 2025 | Data IntegrationLarge Language Model | —Unverified | 0 |
| Load Balancing Mixture of Experts with Similarity Preserving Routers | Jun 16, 2025 | Mixture-of-Experts | —Unverified | 0 |
| EAQuant: Enhancing Post-Training Quantization for MoE Models via Expert-Aware Optimization | Jun 16, 2025 | Mixture-of-ExpertsModel Compression | CodeCode Available | 0 |
| Serving Large Language Models on Huawei CloudMatrix384 | Jun 15, 2025 | Mixture-of-ExpertsQuantization | —Unverified | 0 |
| Optimus-3: Towards Generalist Multimodal Minecraft Agents with Scalable Task Experts | Jun 12, 2025 | DiversityMinecraft | —Unverified | 0 |
| GigaChat Family: Efficient Russian Language Modeling Through Mixture of Experts Architecture | Jun 11, 2025 | Language ModelingLanguage Modelling | —Unverified | 0 |
| MedMoE: Modality-Specialized Mixture of Experts for Medical Vision-Language Understanding | Jun 10, 2025 | DiagnosticMixture-of-Experts | —Unverified | 0 |
| A Two-Phase Deep Learning Framework for Adaptive Time-Stepping in High-Speed Flow Modeling | Jun 9, 2025 | Mixture-of-Experts | —Unverified | 0 |
| MoE-MLoRA for Multi-Domain CTR Prediction: Efficient Adaptation with Expert Specialization | Jun 9, 2025 | Click-Through Rate PredictionDiversity | CodeCode Available | 0 |
| STAMImputer: Spatio-Temporal Attention MoE for Traffic Data Imputation | Jun 9, 2025 | Graph AttentionImputation | CodeCode Available | 0 |
| MIRA: Medical Time Series Foundation Model for Real-World Health Data | Jun 9, 2025 | EthicsMissing Values | —Unverified | 0 |
| M2Restore: Mixture-of-Experts-based Mamba-CNN Fusion Framework for All-in-One Image Restoration | Jun 9, 2025 | AllImage Restoration | —Unverified | 0 |
| MoE-GPS: Guidlines for Prediction Strategy for Dynamic Expert Duplication in MoE Load Balancing | Jun 9, 2025 | GPUMixture-of-Experts | —Unverified | 0 |
| Breaking Data Silos: Towards Open and Scalable Mobility Foundation Models via Generative Continual Learning | Jun 7, 2025 | Continual LearningFederated Learning | —Unverified | 0 |
| SMAR: Soft Modality-Aware Routing Strategy for MoE-based Multimodal Large Language Models Preserving Language Capabilities | Jun 6, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Lifelong Evolution: Collaborative Learning between Large and Small Language Models for Continuous Emergent Fake News Detection | Jun 5, 2025 | Fake News Detectionknowledge editing | —Unverified | 0 |
| Brain-Like Processing Pathways Form in Models With Heterogeneous Experts | Jun 3, 2025 | FormMixture-of-Experts | —Unverified | 0 |
| Enhancing Multimodal Continual Instruction Tuning with BranchLoRA | May 31, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Decoding Knowledge Attribution in Mixture-of-Experts: A Framework of Basic-Refinement Collaboration and Efficiency Analysis | May 30, 2025 | BlockingMixture-of-Experts | —Unverified | 0 |
| Mixture-of-Experts for Personalized and Semantic-Aware Next Location Prediction | May 30, 2025 | Domain GeneralizationMixture-of-Experts | —Unverified | 0 |
| GradPower: Powering Gradients for Faster Language Model Pre-Training | May 30, 2025 | Language ModelingLanguage Modelling | —Unverified | 0 |
| On the Expressive Power of Mixture-of-Experts for Structured Complex Tasks | May 30, 2025 | Mixture-of-Experts | —Unverified | 0 |
| From Knowledge to Noise: CTIM-Rover and the Pitfalls of Episodic Memory in Software Engineering Agents | May 29, 2025 | AI AgentMixture-of-Experts | CodeCode Available | 0 |
| Two Is Better Than One: Rotations Scale LoRAs | May 29, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Revisiting Uncertainty Estimation and Calibration of Large Language Models | May 29, 2025 | Mixture-of-ExpertsMMLU | —Unverified | 0 |
| Noise-Robustness Through Noise: Asymmetric LoRA Adaption with Poisoning Expert | May 29, 2025 | Mixture-of-Expertsparameter-efficient fine-tuning | —Unverified | 0 |
| A Survey of Generative Categories and Techniques in Multimodal Large Language Models | May 29, 2025 | Mixture-of-ExpertsSelf-Supervised Learning | —Unverified | 0 |
| Point-MoE: Towards Cross-Domain Generalization in 3D Semantic Segmentation via Mixture-of-Experts | May 29, 2025 | 3D Semantic SegmentationDomain Generalization | —Unverified | 0 |
| A Human-Centric Approach to Explainable AI for Personalized Education | May 28, 2025 | Autonomous DrivingMixture-of-Experts | CodeCode Available | 0 |
| Advancing Expert Specialization for Better MoE | May 28, 2025 | Mixture-of-Experts | —Unverified | 0 |
| EvoMoE: Expert Evolution in Mixture of Experts for Multimodal Large Language Models | May 28, 2025 | Mixture-of-ExpertsMME | —Unverified | 0 |
| ForceVLA: Enhancing VLA Models with a Force-aware MoE for Contact-rich Manipulation | May 28, 2025 | Contact-rich ManipulationMixture-of-Experts | —Unverified | 0 |
| MoE-Gyro: Self-Supervised Over-Range Reconstruction and Denoising for MEMS Gyroscopes | May 27, 2025 | BenchmarkingDenoising | —Unverified | 0 |
| Mosaic: Data-Free Knowledge Distillation via Mixture-of-Experts for Heterogeneous Distributed Environments | May 26, 2025 | Data-free Knowledge DistillationFederated Learning | CodeCode Available | 0 |
| MoESD: Unveil Speculative Decoding's Potential for Accelerating Sparse MoE | May 26, 2025 | Mixture-of-Experts | —Unverified | 0 |
| NEXT: Multi-Grained Mixture of Experts via Text-Modulation for Multi-Modal Object Re-ID | May 26, 2025 | AttributeCaption Generation | —Unverified | 0 |
| Rethinking Gating Mechanism in Sparse MoE: Handling Arbitrary Modality Inputs with Confidence-Guided Gate | May 26, 2025 | ImputationMixture-of-Experts | CodeCode Available | 0 |
| RankLLM: A Python Package for Reranking with LLMs | May 25, 2025 | Mixture-of-ExpertsRAG | —Unverified | 0 |
| Integrating Dynamical Systems Learning with Foundational Models: A Meta-Evolutionary AI Framework for Clinical Trials | May 25, 2025 | Evolutionary AlgorithmsLarge Language Model | —Unverified | 0 |
| μ-MoE: Test-Time Pruning as Micro-Grained Mixture-of-Experts | May 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| On Minimax Estimation of Parameters in Softmax-Contaminated Mixture of Experts | May 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Guiding the Experts: Semantic Priors for Efficient and Focused MoE Routing | May 24, 2025 | Mixture-of-Experts | CodeCode Available | 0 |
| Mod-Adapter: Tuning-Free and Versatile Multi-concept Personalization via Modulation Adapter | May 24, 2025 | Image GenerationMixture-of-Experts | —Unverified | 0 |
| TrajMoE: Spatially-Aware Mixture of Experts for Unified Human Mobility Modeling | May 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| EvidenceMoE: A Physics-Guided Mixture-of-Experts with Evidential Critics for Advancing Fluorescence Light Detection and Ranging in Scattering Media | May 23, 2025 | Depth EstimationMixture-of-Experts | —Unverified | 0 |
| DualComp: End-to-End Learning of a Unified Dual-Modality Lossless Compressor | May 22, 2025 | Mixture-of-Experts | —Unverified | 0 |
| DriveMoE: Mixture-of-Experts for Vision-Language-Action Model in End-to-End Autonomous Driving | May 22, 2025 | Autonomous DrivingBench2Drive | —Unverified | 0 |
| Time Tracker: Mixture-of-Experts-Enhanced Foundation Time Series Forecasting Model with Decoupled Training Pipelines | May 21, 2025 | Graph LearningMixture-of-Experts | —Unverified | 0 |