| MoFE: Mixture of Frozen Experts Architecture | Mar 9, 2025 | Mixture-of-Expertsparameter-efficient fine-tuning | —Unverified | 0 | 0 |
| MoHAVE: Mixture of Hierarchical Audio-Visual Experts for Robust Speech Recognition | Feb 11, 2025 | Audio-Visual Speech RecognitionComputational Efficiency | —Unverified | 0 | 0 |
| MoIN: Mixture of Introvert Experts to Upcycle an LLM | Oct 13, 2024 | GPULanguage Modeling | —Unverified | 0 | 0 |
| MoLe-VLA: Dynamic Layer-skipping Vision Language Action Model via Mixture-of-Layers for Efficient Robot Manipulation | Mar 26, 2025 | Knowledge DistillationMixture-of-Experts | —Unverified | 0 | 0 |
| MolGraph-xLSTM: A graph-based dual-level xLSTM framework with multi-head mixture-of-experts for enhanced molecular representation and interpretability | Jan 30, 2025 | Drug DiscoveryMixture-of-Experts | —Unverified | 0 | 0 |
| MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts | Jul 31, 2024 | Causal InferenceLanguage Modelling | —Unverified | 0 | 0 |
| MoMQ: Mixture-of-Experts Enhances Multi-Dialect Query Generation across Relational and Non-Relational Databases | Oct 24, 2024 | Mixture-of-Experts | —Unverified | 0 | 0 |
| MoNDE: Mixture of Near-Data Experts for Large-Scale Sparse Models | May 29, 2024 | DecoderGPU | —Unverified | 0 | 0 |
| Mono-InternVL: Pushing the Boundaries of Monolithic Multimodal Large Language Models with Endogenous Visual Pre-training | Oct 10, 2024 | Mixture-of-ExpertsVisual Question Answering | —Unverified | 0 | 0 |
| MoPEFT: A Mixture-of-PEFTs for the Segment Anything Model | May 1, 2024 | Mixture-of-Expertsparameter-efficient fine-tuning | —Unverified | 0 | 0 |
| MoRAL: MoE Augmented LoRA for LLMs' Lifelong Learning | Feb 17, 2024 | Lifelong learningMixture-of-Experts | —Unverified | 0 | 0 |
| MoRE: Unlocking Scalability in Reinforcement Learning for Quadruped Vision-Language-Action Models | Mar 11, 2025 | Large Language ModelMixture-of-Experts | —Unverified | 0 | 0 |