| MoESD: Unveil Speculative Decoding's Potential for Accelerating Sparse MoE | May 26, 2025 | Mixture-of-Experts | —Unverified | 0 |
| WINA: Weight Informed Neuron Activation for Accelerating Large Language Model Inference | May 26, 2025 | Language ModelingLanguage Modelling | CodeCode Available | 2 |
| FLAME-MoE: A Transparent End-to-End Research Platform for Mixture-of-Experts Language Models | May 26, 2025 | Mixture-of-Experts | CodeCode Available | 1 |
| NEXT: Multi-Grained Mixture of Experts via Text-Modulation for Multi-Modal Object Re-ID | May 26, 2025 | AttributeCaption Generation | —Unverified | 0 |
| Mosaic: Data-Free Knowledge Distillation via Mixture-of-Experts for Heterogeneous Distributed Environments | May 26, 2025 | Data-free Knowledge DistillationFederated Learning | CodeCode Available | 0 |
| Integrating Dynamical Systems Learning with Foundational Models: A Meta-Evolutionary AI Framework for Clinical Trials | May 25, 2025 | Evolutionary AlgorithmsLarge Language Model | —Unverified | 0 |
| RankLLM: A Python Package for Reranking with LLMs | May 25, 2025 | Mixture-of-ExpertsRAG | CodeCode Available | 0 |
| I2MoE: Interpretable Multimodal Interaction-aware Mixture-of-Experts | May 25, 2025 | Mixture-of-Expertsmultimodal interaction | CodeCode Available | 2 |
| ThanoRA: Task Heterogeneity-Aware Multi-Task Low-Rank Adaptation | May 24, 2025 | Mixture-of-Experts | CodeCode Available | 1 |
| On Minimax Estimation of Parameters in Softmax-Contaminated Mixture of Experts | May 24, 2025 | Mixture-of-Experts | —Unverified | 0 |