| AT-MoE: Adaptive Task-planning Mixture of Experts via LoRA Approach | Oct 12, 2024 | Mixture-of-ExpertsTask Planning | —Unverified | 0 |
| HDformer: A Higher Dimensional Transformer for Diabetes Detection Utilizing Long Range Vascular Signals | Mar 17, 2023 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| HeterMoE: Efficient Training of Mixture-of-Experts Models on Heterogeneous GPUs | Apr 4, 2025 | GPUMixture-of-Experts | —Unverified | 0 |
| DADNN: Multi-Scene CTR Prediction via Domain-Aware Deep Neural Network | Nov 24, 2020 | Click-Through Rate PredictionMixture-of-Experts | —Unverified | 0 |
| D^2MoE: Dual Routing and Dynamic Scheduling for Efficient On-Device MoE-based LLM Serving | Apr 17, 2025 | Mixture-of-ExpertsModel Compression | —Unverified | 0 |
| A Theoretical View on Sparsely Activated Networks | Aug 8, 2022 | Mixture-of-Experts | —Unverified | 0 |
| A Large-scale Medical Visual Task Adaptation Benchmark | Apr 19, 2024 | Mixture-of-Experts | —Unverified | 0 |
| HMDN: Hierarchical Multi-Distribution Network for Click-Through Rate Prediction | Aug 2, 2024 | Click-Through Rate PredictionMixture-of-Experts | —Unverified | 0 |
| CSAOT: Cooperative Multi-Agent System for Active Object Tracking | Jan 23, 2025 | Autonomous NavigationDeep Reinforcement Learning | —Unverified | 0 |
| Cross-Topic Rumor Detection using Topic-Mixtures | Apr 1, 2021 | Mixture-of-Experts | —Unverified | 0 |
| A Survey on Model MoErging: Recycling and Routing Among Specialized Experts for Collaborative Learning | Aug 13, 2024 | Mixture-of-ExpertsSurvey | —Unverified | 0 |
| AIREX: Neural Network-based Approach for Air Quality Inference in Unmonitored Cities | Aug 16, 2021 | Air Quality InferenceMixture-of-Experts | —Unverified | 0 |
| PMoE: Progressive Mixture of Experts with Asymmetric Transformer for Continual Learning | Jul 31, 2024 | Continual LearningGeneral Knowledge | —Unverified | 0 |
| Hard Mixtures of Experts for Large Scale Weakly Supervised Vision | Apr 20, 2017 | GPUMixture-of-Experts | —Unverified | 0 |
| Heuristic-Informed Mixture of Experts for Link Prediction in Multilayer Networks | Jan 29, 2025 | Link PredictionMixture-of-Experts | —Unverified | 0 |
| Agent4Ranking: Semantic Robust Ranking via Personalized Query Rewriting Using Multi-agent LLM | Dec 24, 2023 | Mixture-of-Experts | —Unverified | 0 |
| Adapted-MoE: Mixture of Experts with Test-Time Adaption for Anomaly Detection | Sep 9, 2024 | Anomaly DetectionMixture-of-Experts | —Unverified | 0 |
| CoSMoEs: Compact Sparse Mixture of Experts | Feb 28, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Correlative and Discriminative Label Grouping for Multi-Label Visual Prompt Tuning | Jan 1, 2025 | image-classificationImage Classification | —Unverified | 0 |
| GRIN: GRadient-INformed MoE | Sep 18, 2024 | HellaSwagHumanEval | —Unverified | 0 |
| Core-Periphery Principle Guided State Space Model for Functional Connectome Classification | Mar 18, 2025 | Functional ConnectivityMamba | —Unverified | 0 |
| Coordination with Humans via Strategy Matching | Oct 27, 2022 | Mixture-of-Experts | —Unverified | 0 |
| A Survey on Dynamic Neural Networks for Natural Language Processing | Feb 15, 2022 | Dynamic neural networksMixture-of-Experts | —Unverified | 0 |
| Convolutional Neural Networks and Mixture of Experts for Intrusion Detection in 5G Networks and beyond | Dec 4, 2024 | Intrusion DetectionMixture-of-Experts | —Unverified | 0 |
| Convergence Rates for Softmax Gating Mixture of Experts | Mar 5, 2025 | Mixture-of-Expertsparameter estimation | —Unverified | 0 |
| Astrea: A MOE-based Visual Understanding Model with Progressive Alignment | Mar 12, 2025 | Contrastive LearningCross-Modal Retrieval | —Unverified | 0 |
| HAECcity: Open-Vocabulary Scene Understanding of City-Scale Point Clouds with Superpoint Graph Clustering | Apr 18, 2025 | ClusteringGraph Clustering | —Unverified | 0 |
| Continual Traffic Forecasting via Mixture of Experts | Jun 5, 2024 | Continual LearningMixture-of-Experts | —Unverified | 0 |
| Improving Transformer Performance for French Clinical Notes Classification Using Mixture of Experts on a Limited Dataset | Mar 22, 2023 | Mixture-of-Expertstext-classification | —Unverified | 0 |
| Continual Pre-training of MoEs: How robust is your router? | Mar 6, 2025 | DecoderMixture-of-Experts | —Unverified | 0 |
| Continual Learning Using Task Conditional Neural Networks | Sep 29, 2021 | Continual LearningMixture-of-Experts | —Unverified | 0 |
| A General Theory for Softmax Gating Multinomial Logistic Mixture of Experts | Oct 22, 2023 | Density EstimationMixture-of-Experts | —Unverified | 0 |
| A Generalist Cross-Domain Molecular Learning Framework for Structure-Based Drug Discovery | Mar 6, 2025 | DenoisingDrug Discovery | —Unverified | 0 |
| ContextWIN: Whittle Index Based Mixture-of-Experts Neural Model For Restless Bandits Via Deep RL | Oct 13, 2024 | Decision MakingMixture-of-Experts | —Unverified | 0 |
| Contextual Policy Transfer in Reinforcement Learning Domains via Deep Mixtures-of-Experts | Feb 29, 2020 | Mixture-of-ExpertsOpenAI Gym | —Unverified | 0 |
| A Simple Architecture for Enterprise Large Language Model Applications based on Role based security and Clearance Levels using Retrieval-Augmented Generation or Mixture of Experts | Jul 9, 2024 | Language ModelingLanguage Modelling | —Unverified | 0 |
| Contextual Mixture of Experts: Integrating Knowledge into Predictive Modeling | Nov 1, 2022 | Mixture-of-Experts | —Unverified | 0 |
| ConstitutionalExperts: Training a Mixture of Principle-based Prompts | Mar 7, 2024 | Mixture-of-Experts | —Unverified | 0 |
| A similarity-based Bayesian mixture-of-experts model | Dec 3, 2020 | Mixture-of-Expertsmodel | —Unverified | 0 |
| Half-Space Feature Learning in Neural Networks | Apr 5, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Connector-S: A Survey of Connectors in Multi-modal Large Language Models | Feb 17, 2025 | Mixture-of-ExpertsSurvey | —Unverified | 0 |
| Configurable Foundation Models: Building LLMs from a Modular Perspective | Sep 4, 2024 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| 3D-MoE: A Mixture-of-Experts Multi-modal LLM for 3D Vision and Pose Diffusion via Rectified Flow | Jan 28, 2025 | Instruction FollowingMixture-of-Experts | —Unverified | 0 |
| Conditional computation in neural networks: principles and research trends | Mar 12, 2024 | Mixture-of-Expertsscientific discovery | —Unverified | 0 |
| On DeepSeekMoE: Statistical Benefits of Shared Experts and Normalized Sigmoid Gating | May 16, 2025 | Language ModelingLanguage Modelling | —Unverified | 0 |
| On the Adaptation to Concept Drift for CTR Prediction | Apr 1, 2022 | Click-Through Rate PredictionIncremental Learning | —Unverified | 0 |
| A Review of Sparse Expert Models in Deep Learning | Sep 4, 2022 | Deep LearningMixture-of-Experts | —Unverified | 0 |
| Complexity Experts are Task-Discriminative Learners for Any Image Restoration | Nov 27, 2024 | AttributeBlind All-in-One Image Restoration | —Unverified | 0 |
| Filtered not Mixed: Stochastic Filtering-Based Online Gating for Mixture of Large Language Models | Jun 5, 2024 | Mixture-of-ExpertsTime Series | —Unverified | 0 |
| A Review of DeepSeek Models' Key Innovative Techniques | Mar 14, 2025 | Mixture-of-Expertsreinforcement-learning | —Unverified | 0 |