| Every Expert Matters: Towards Effective Knowledge Distillation for Mixture-of-Experts Language Models | Feb 18, 2025 | Knowledge DistillationMixture-of-Experts | —Unverified | 0 |
| EVE: Efficient Vision-Language Pre-training with Masked Prediction and Modality-Aware MoE | Aug 23, 2023 | Image-text matchingImage-text Retrieval | —Unverified | 0 |
| Channel Gain Cartography via Mixture of Experts | Dec 8, 2020 | Mixture-of-Experts | —Unverified | 0 |
| EVA: Mixture-of-Experts Semantic Variant Alignment for Compositional Zero-Shot Learning | Jun 26, 2025 | Compositional Zero-Shot LearningMixture-of-Experts | —Unverified | 0 |
| Evaluating Expert Contributions in a MoE LLM for Quiz-Based Tasks | Feb 24, 2025 | Mixture-of-ExpertsMMLU | —Unverified | 0 |
| Changing Model Behavior at Test-Time Using Reinforcement Learning | Feb 24, 2017 | BIG-bench Machine LearningMixture-of-Experts | —Unverified | 0 |
| ADMoE: Anomaly Detection with Mixture-of-Experts from Noisy Labels | Aug 24, 2022 | Anomaly DetectionMixture-of-Experts | —Unverified | 0 |
| Modular Action Concept Grounding in Semantic Video Prediction | Nov 23, 2020 | Action RecognitionMixture-of-Experts | —Unverified | 0 |
| EPS-MoE: Expert Pipeline Scheduler for Cost-Efficient MoE Inference | Oct 16, 2024 | Computational EfficiencyLarge Language Model | —Unverified | 0 |
| LoRA-Switch: Boosting the Efficiency of Dynamic LLM Adapters via System-Algorithm Co-design | May 28, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Ensemble Learning for Large Language Models in Text and Code Generation: A Survey | Mar 13, 2025 | Code GenerationEnsemble Learning | —Unverified | 0 |
| Non-asymptotic oracle inequalities for the Lasso in high-dimensional mixture of experts | Sep 22, 2020 | feature selectionMixture-of-Experts | —Unverified | 0 |
| Routing in Sparsely-gated Language Models responds to Context | Sep 21, 2024 | DecoderMixture-of-Experts | —Unverified | 0 |
| Lory: Fully Differentiable Mixture-of-Experts for Autoregressive Language Model Pre-training | May 6, 2024 | Language ModelingLanguage Modelling | —Unverified | 0 |
| Low-Rank Mixture-of-Experts for Continual Medical Image Segmentation | Jun 19, 2024 | Continual LearningImage Segmentation | —Unverified | 0 |
| Enhancing the "Immunity" of Mixture-of-Experts Networks for Adversarial Defense | Feb 29, 2024 | Adversarial DefenseAdversarial Robustness | —Unverified | 0 |
| Capacity-Aware Inference: Mitigating the Straggler Effect in Mixture of Experts | Mar 7, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Enhancing Multi-modal Models with Heterogeneous MoE Adapters for Fine-tuning | Mar 26, 2025 | Mixture-of-Expertsparameter-efficient fine-tuning | —Unverified | 0 |
| Enhancing Multimodal Continual Instruction Tuning with BranchLoRA | May 31, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Context-aware Mixture-of-Experts for Unbiased Scene Graph Generation | Aug 15, 2022 | DiversityGraph Generation | —Unverified | 0 |
| An Introduction to the Practical and Theoretical Aspects of Mixture-of-Experts Modeling | Jul 12, 2017 | ClusteringMixture-of-Experts | —Unverified | 0 |
| Enhancing Healthcare Recommendation Systems with a Multimodal LLMs-based MOE Architecture | Dec 16, 2024 | Mixture-of-ExpertsRecommendation Systems | —Unverified | 0 |
| Enhancing Generalization in Sparse Mixture of Experts Models: The Case for Increased Expert Activation in Compositional Tasks | Oct 17, 2024 | Mixture-of-Experts | —Unverified | 0 |
| CAME: Competitively Learning a Mixture-of-Experts Model for First-stage Retrieval | Nov 6, 2023 | Mixture-of-ExpertsRetrieval | —Unverified | 0 |
| An Entailment Tree Generation Approach for Multimodal Multi-Hop Question Answering with Mixture-of-Experts and Iterative Feedback Mechanism | Dec 8, 2024 | Mixture-of-ExpertsMulti-hop Question Answering | —Unverified | 0 |