| Optimal Scaling Laws for Efficiency Gains in a Theoretical Transformer-Augmented Sectional MoE Framework | Mar 26, 2025 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| Enhancing Multi-modal Models with Heterogeneous MoE Adapters for Fine-tuning | Mar 26, 2025 | Mixture-of-Expertsparameter-efficient fine-tuning | —Unverified | 0 |
| M^2CD: A Unified MultiModal Framework for Optical-SAR Change Detection with Mixture of Experts and Self-Distillation | Mar 25, 2025 | Change DetectionDisaster Response | —Unverified | 0 |
| Resilient Sensor Fusion under Adverse Sensor Failures via Multi-Modal Expert Fusion | Mar 25, 2025 | Autonomous DrivingMixture-of-Experts | —Unverified | 0 |
| BiPrompt-SAM: Enhancing Image Segmentation via Explicit Selection between Point and Text Prompts | Mar 25, 2025 | Image SegmentationMixture-of-Experts | —Unverified | 0 |
| SPMTrack: Spatio-Temporal Parameter-Efficient Fine-Tuning with Mixture of Experts for Scalable Visual Tracking | Mar 24, 2025 | Mixture-of-Expertsparameter-efficient fine-tuning | CodeCode Available | 1 |
| Galaxy Walker: Geometry-aware VLMs For Galaxy-scale Understanding | Mar 24, 2025 | Mixture-of-ExpertsMorphology classification | —Unverified | 0 |
| ExpertRAG: Efficient RAG with Mixture of Experts -- Optimizing Context Retrieval for Adaptive LLM Responses | Mar 23, 2025 | Language ModelingLanguage Modelling | —Unverified | 0 |
| Every Sample Matters: Leveraging Mixture-of-Experts and High-Quality Data for Efficient and Accurate Code LLM | Mar 22, 2025 | Code GenerationMixture-of-Experts | —Unverified | 0 |
| Expert Race: A Flexible Routing Strategy for Scaling Diffusion Transformer with Mixture of Experts | Mar 20, 2025 | Mixture-of-Experts | —Unverified | 0 |
| UniCoRN: Latent Diffusion-based Unified Controllable Image Restoration Network across Multiple Degradations | Mar 20, 2025 | Image RestorationMixture-of-Experts | —Unverified | 0 |
| Mixture of Lookup Experts | Mar 20, 2025 | Mixture-of-Experts | CodeCode Available | 2 |
| SemEval-2025 Task 1: AdMIRe -- Advancing Multimodal Idiomaticity Representation | Mar 19, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Leveraging MoE-based Large Language Model for Zero-Shot Multi-Task Semantic Communication | Mar 19, 2025 | Language ModelingLanguage Modelling | —Unverified | 0 |
| Core-Periphery Principle Guided State Space Model for Functional Connectome Classification | Mar 18, 2025 | Functional ConnectivityMamba | —Unverified | 0 |
| MAST-Pro: Dynamic Mixture-of-Experts for Adaptive Segmentation of Pan-Tumors with Knowledge-Driven Prompts | Mar 18, 2025 | Mixture-of-Expertsparameter-efficient fine-tuning | —Unverified | 0 |
| Fast filtering of non-Gaussian models using Amortized Optimal Transport Maps | Mar 16, 2025 | Mixture-of-Experts | CodeCode Available | 0 |
| Adaptive Mixture of Low-Rank Experts for Robust Audio Spoofing Detection | Mar 15, 2025 | Mixture-of-Experts | —Unverified | 0 |
| A Review of DeepSeek Models' Key Innovative Techniques | Mar 14, 2025 | Mixture-of-Expertsreinforcement-learning | —Unverified | 0 |
| MoLEx: Mixture of Layer Experts for Finetuning with Sparse Upcycling | Mar 14, 2025 | Mixture-of-Expertsparameter-efficient fine-tuning | CodeCode Available | 0 |
| Ensemble Learning for Large Language Models in Text and Code Generation: A Survey | Mar 13, 2025 | Code GenerationEnsemble Learning | —Unverified | 0 |
| StableFusion: Continual Video Retrieval via Frame Adaptation | Mar 13, 2025 | Continual LearningMixture-of-Experts | CodeCode Available | 1 |
| dFLMoE: Decentralized Federated Learning via Mixture of Experts for Medical Data Analysis | Mar 13, 2025 | Federated LearningMixture-of-Experts | —Unverified | 0 |
| Samoyeds: Accelerating MoE Models with Structured Sparsity Leveraging Sparse Tensor Cores | Mar 13, 2025 | Mixture-of-Experts | CodeCode Available | 1 |
| Astrea: A MOE-based Visual Understanding Model with Progressive Alignment | Mar 12, 2025 | Contrastive LearningCross-Modal Retrieval | —Unverified | 0 |
| FaVChat: Unlocking Fine-Grained Facail Video Understanding with Multimodal Large Language Models | Mar 12, 2025 | Mixture-of-ExpertsQuestion Answering | —Unverified | 0 |
| Towards Robust Multimodal Representation: A Unified Approach with Adaptive Experts and Alignment | Mar 12, 2025 | Contrastive LearningDecision Making | CodeCode Available | 0 |
| Double-Stage Feature-Level Clustering-Based Mixture of Experts Framework | Mar 12, 2025 | ClusteringDiversity | —Unverified | 0 |
| Priority-Aware Preemptive Scheduling for Mixed-Priority Workloads in MoE Inference | Mar 12, 2025 | BlockingGPU | —Unverified | 0 |
| Automatic Operator-level Parallelism Planning for Distributed Deep Learning -- A Mixed-Integer Programming Approach | Mar 12, 2025 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| MoRE: Unlocking Scalability in Reinforcement Learning for Quadruped Vision-Language-Action Models | Mar 11, 2025 | Large Language ModelMixture-of-Experts | —Unverified | 0 |
| UniF^2ace: Fine-grained Face Understanding and Generation with Unified Multimodal Models | Mar 11, 2025 | AttributeMixture-of-Experts | —Unverified | 0 |
| MoE-Loco: Mixture of Experts for Multitask Locomotion | Mar 11, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Accelerating MoE Model Inference with Expert Sharding | Mar 11, 2025 | DecoderGPU | —Unverified | 0 |
| GM-MoE: Low-Light Enhancement with Gated-Mechanism Mixture-of-Experts | Mar 10, 2025 | 3D ReconstructionAutonomous Driving | —Unverified | 0 |
| A Comprehensive Survey of Mixture-of-Experts: Algorithms, Theory, and Applications | Mar 10, 2025 | Continual LearningMeta-Learning | CodeCode Available | 9 |
| eMoE: Task-aware Memory Efficient Mixture-of-Experts-Based (MoE) Model Inference | Mar 10, 2025 | Mixture-of-ExpertsScheduling | —Unverified | 0 |
| ResMoE: Space-efficient Compression of Mixture of Experts LLMs via Residual Restoration | Mar 10, 2025 | Mixture-of-Experts | CodeCode Available | 0 |
| Swift Hydra: Self-Reinforcing Generative Framework for Anomaly Detection with Multiple Mamba Models | Mar 9, 2025 | Anomaly DetectionMamba | CodeCode Available | 0 |
| MoFE: Mixture of Frozen Experts Architecture | Mar 9, 2025 | Mixture-of-Expertsparameter-efficient fine-tuning | —Unverified | 0 |
| MANDARIN: Mixture-of-Experts Framework for Dynamic Delirium and Coma Prediction in ICU Patients: Development and Validation of an Acute Brain Dysfunction Prediction Model | Mar 8, 2025 | Mixture-of-Experts | —Unverified | 0 |
| A Novel Trustworthy Video Summarization Algorithm Through a Mixture of LoRA Experts | Mar 8, 2025 | Mixture-of-ExpertsVideo Summarization | —Unverified | 0 |
| MoEMoE: Question Guided Dense and Scalable Sparse Mixture-of-Expert for Multi-source Multi-modal Answering | Mar 8, 2025 | Answer GenerationMixture-of-Experts | —Unverified | 0 |
| Capacity-Aware Inference: Mitigating the Straggler Effect in Mixture of Experts | Mar 7, 2025 | Mixture-of-Experts | —Unverified | 0 |
| FMT:A Multimodal Pneumonia Detection Model Based on Stacking MOE Framework | Mar 7, 2025 | DiagnosticMedical Image Analysis | —Unverified | 0 |
| Every FLOP Counts: Scaling a 300B Mixture-of-Experts LING LLM without Premium GPUs | Mar 7, 2025 | Knowledge GraphsMixture-of-Experts | —Unverified | 0 |
| Symbolic Mixture-of-Experts: Adaptive Skill-based Routing for Heterogeneous Reasoning | Mar 7, 2025 | GPUMath | —Unverified | 0 |
| Linear-MoE: Linear Sequence Modeling Meets Mixture-of-Experts | Mar 7, 2025 | Mixture-of-ExpertsState Space Models | CodeCode Available | 2 |
| Continual Pre-training of MoEs: How robust is your router? | Mar 6, 2025 | DecoderMixture-of-Experts | —Unverified | 0 |
| TS-RAG: Retrieval-Augmented Generation based Time Series Foundation Models are Stronger Zero-Shot Forecaster | Mar 6, 2025 | Domain AdaptationMixture-of-Experts | —Unverified | 0 |