| Integrating Dynamical Systems Learning with Foundational Models: A Meta-Evolutionary AI Framework for Clinical Trials | May 25, 2025 | Evolutionary AlgorithmsLarge Language Model | —Unverified | 0 |
| RankLLM: A Python Package for Reranking with LLMs | May 25, 2025 | Mixture-of-ExpertsRAG | CodeCode Available | 0 |
| I2MoE: Interpretable Multimodal Interaction-aware Mixture-of-Experts | May 25, 2025 | Mixture-of-Expertsmultimodal interaction | CodeCode Available | 2 |
| On Minimax Estimation of Parameters in Softmax-Contaminated Mixture of Experts | May 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| TrajMoE: Spatially-Aware Mixture of Experts for Unified Human Mobility Modeling | May 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| ThanoRA: Task Heterogeneity-Aware Multi-Task Low-Rank Adaptation | May 24, 2025 | Mixture-of-Experts | CodeCode Available | 1 |
| μ-MoE: Test-Time Pruning as Micro-Grained Mixture-of-Experts | May 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Mod-Adapter: Tuning-Free and Versatile Multi-concept Personalization via Modulation Adapter | May 24, 2025 | Image GenerationMixture-of-Experts | —Unverified | 0 |
| Guiding the Experts: Semantic Priors for Efficient and Focused MoE Routing | May 24, 2025 | Mixture-of-Experts | CodeCode Available | 0 |
| EvidenceMoE: A Physics-Guided Mixture-of-Experts with Evidential Critics for Advancing Fluorescence Light Detection and Ranging in Scattering Media | May 23, 2025 | Depth EstimationMixture-of-Experts | —Unverified | 0 |
| JanusDNA: A Powerful Bi-directional Hybrid DNA Foundation Model | May 22, 2025 | GPULong-range modeling | CodeCode Available | 1 |
| DualComp: End-to-End Learning of a Unified Dual-Modality Lossless Compressor | May 22, 2025 | Mixture-of-Experts | —Unverified | 0 |
| DriveMoE: Mixture-of-Experts for Vision-Language-Action Model in End-to-End Autonomous Driving | May 22, 2025 | Autonomous DrivingBench2Drive | —Unverified | 0 |
| Not All Models Suit Expert Offloading: On Local Routing Consistency of Mixture-of-Expert Models | May 21, 2025 | AllCPU | CodeCode Available | 0 |
| CoLA: Collaborative Low-Rank Adaptation | May 21, 2025 | CoLAMixture-of-Experts | CodeCode Available | 0 |
| Time Tracker: Mixture-of-Experts-Enhanced Foundation Time Series Forecasting Model with Decoupled Training Pipelines | May 21, 2025 | Graph LearningMixture-of-Experts | —Unverified | 0 |
| MoRE-Brain: Routed Mixture of Experts for Interpretable and Generalizable Cross-Subject fMRI Visual Decoding | May 21, 2025 | Mixture-of-Experts | CodeCode Available | 0 |
| Hunyuan-TurboS: Advancing Large Language Models through Mamba-Transformer Synergy and Adaptive Chain-of-Thought | May 21, 2025 | ChatbotInstruction Following | —Unverified | 0 |
| Efficient Data Driven Mixture-of-Expert Extraction from Trained Networks | May 21, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Multimodal Cultural Safety: Evaluation Frameworks and Alignment Strategies | May 20, 2025 | Mixture-of-Experts | CodeCode Available | 0 |
| Balanced and Elastic End-to-end Training of Dynamic LLMs | May 20, 2025 | GPUMixture-of-Experts | —Unverified | 0 |
| Multimodal Mixture of Low-Rank Experts for Sentiment Analysis and Emotion Recognition | May 20, 2025 | Emotion RecognitionMixture-of-Experts | —Unverified | 0 |
| THOR-MoE: Hierarchical Task-Guided and Context-Responsive Routing for Neural Machine Translation | May 20, 2025 | Machine TranslationMixture-of-Experts | —Unverified | 0 |
| Two Experts Are All You Need for Steering Thinking: Reinforcing Cognitive Effort in MoE Reasoning Models Without Additional Training | May 20, 2025 | AllDomain Generalization | —Unverified | 0 |
| FuxiMT: Sparsifying Large Language Models for Chinese-Centric Multilingual Machine Translation | May 20, 2025 | Language ModelingLanguage Modelling | —Unverified | 0 |