| Omni-SMoLA: Boosting Generalist Multimodal Models with Soft Mixture of Low-rank Experts | Dec 1, 2023 | Chart Question AnsweringDocument AI | —Unverified | 0 |
| OMoE: Diversifying Mixture of Low-Rank Adaptation by Orthogonal Finetuning | Jan 17, 2025 | Computational EfficiencyDiversity | —Unverified | 0 |
| On component interactions in two-stage recommender systems | Jun 28, 2021 | Mixture-of-ExpertsRecommendation Systems | —Unverified | 0 |
| SkillNet-NLU: A Sparsely Activated Model for General-Purpose Natural Language Understanding | Mar 7, 2022 | Language ModellingMasked Language Modeling | —Unverified | 0 |
| OneRec: Unifying Retrieve and Rank with Generative Recommender and Iterative Preference Alignment | Feb 26, 2025 | Mixture-of-ExpertsRecommendation Systems | —Unverified | 0 |
| One Student Knows All Experts Know: From Sparse to Dense | Jan 26, 2022 | AllKnowledge Distillation | —Unverified | 0 |
| On Expert Estimation in Hierarchical Mixture of Experts: Beyond Softmax Gating Functions | Oct 3, 2024 | image-classificationImage Classification | —Unverified | 0 |
| On Least Square Estimation in Softmax Gating Mixture of Experts | Feb 5, 2024 | Mixture-of-Experts | —Unverified | 0 |
| On Minimax Estimation of Parameters in Softmax-Contaminated Mixture of Experts | May 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| On Multi-objective Policy Optimization as a Tool for Reinforcement Learning: Case Studies in Offline RL and Finetuning | Jun 15, 2021 | Deep Reinforcement LearningMixture-of-Experts | —Unverified | 0 |
| On Parameter Estimation in Deviated Gaussian Mixture of Experts | Feb 7, 2024 | Mixture-of-Expertsparameter estimation | —Unverified | 0 |
| On the Adversarial Robustness of Mixture of Experts | Oct 19, 2022 | Adversarial RobustnessMixture-of-Experts | —Unverified | 0 |
| On the effectiveness of discrete representations in sparse mixture of experts | Nov 28, 2024 | Mixture-of-ExpertsQuantization | —Unverified | 0 |
| On the Expressive Power of Mixture-of-Experts for Structured Complex Tasks | May 30, 2025 | Mixture-of-Experts | —Unverified | 0 |
| On the Functional Equivalence of TSK Fuzzy Systems to Neural Networks, Mixture of Experts, CART, and Stacking Ensemble Regression | Mar 25, 2019 | BIG-bench Machine LearningMixture-of-Experts | —Unverified | 0 |
| On the Representation Collapse of Sparse Mixture of Experts | Apr 20, 2022 | ClusteringLanguage Modeling | —Unverified | 0 |
| On the Risk of Evidence Pollution for Malicious Social Text Detection in the Era of LLMs | Oct 16, 2024 | Mixture-of-ExpertsText Detection | —Unverified | 0 |
| Opportunistic Osteoporosis Diagnosis via Texture-Preserving Self-Supervision, Mixture of Experts and Multi-Task Integration | Jun 25, 2025 | Clinical KnowledgeComputed Tomography (CT) | —Unverified | 0 |
| Optimal Scaling Laws for Efficiency Gains in a Theoretical Transformer-Augmented Sectional MoE Framework | Mar 26, 2025 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| Optimizing 6G Integrated Sensing and Communications (ISAC) via Expert Networks | Jun 1, 2024 | ISACMixture-of-Experts | —Unverified | 0 |
| Optimizing Distributed Deployment of Mixture-of-Experts Model Inference in Serverless Computing | Jan 9, 2025 | Bayesian OptimizationCPU | —Unverified | 0 |
| Optimizing LLMs for Resource-Constrained Environments: A Survey of Model Compression Techniques | May 5, 2025 | Knowledge DistillationMixture-of-Experts | —Unverified | 0 |
| Optimizing Mixture-of-Experts Inference Time Combining Model Deployment and Communication Scheduling | Oct 22, 2024 | AllGPU | —Unverified | 0 |
| Optimizing Mixture of Experts using Dynamic Recompilations | May 4, 2022 | Mixture-of-Experts | —Unverified | 0 |
| Optimizing Robustness and Accuracy in Mixture of Experts: A Dual-Model Approach | Feb 5, 2025 | Adversarial RobustnessMixture-of-Experts | —Unverified | 0 |
| Optimus-3: Towards Generalist Multimodal Minecraft Agents with Scalable Task Experts | Jun 12, 2025 | DiversityMinecraft | —Unverified | 0 |
| P-Adapters: Robustly Extracting Factual Information from Language Models with Diverse Prompts | Oct 14, 2021 | Mixture-of-ExpertsNatural Language Queries | —Unverified | 0 |
| Pangu Ultra MoE: How to Train Your Big MoE on Ascend NPUs | May 7, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Parameter-Efficient Conformers via Sharing Sparsely-Gated Experts for End-to-End Speech Recognition | Sep 17, 2022 | Knowledge DistillationMixture-of-Experts | —Unverified | 0 |
| Parameter-Efficient Quantized Mixture-of-Experts Meets Vision-Language Instruction Tuning for Semiconductor Electron Micrograph Analysis | Aug 27, 2024 | Instruction FollowingLanguage Modeling | —Unverified | 0 |
| Parameters vs FLOPs: Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models | Jan 21, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Parameter-varying neural ordinary differential equations with partition-of-unity networks | Oct 1, 2022 | Mixture-of-ExpertsUnity | —Unverified | 0 |
| MoC-System: Efficient Fault Tolerance for Sparse Mixture-of-Experts Model Training | Aug 8, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Part-Of-Speech Sensitivity of Routers in Mixture of Experts Models | Dec 22, 2024 | Mixture-of-ExpertsPOS | —Unverified | 0 |
| PEMT: Multi-Task Correlation Guided Mixture-of-Experts Enables Parameter-Efficient Transfer Learning | Feb 23, 2024 | Mixture-of-Expertsparameter-efficient fine-tuning | —Unverified | 0 |
| Perception-Informed Neural Networks: Beyond Physics-Informed Neural Networks | May 2, 2025 | Mixture-of-Experts | —Unverified | 0 |
| PERFT: Parameter-Efficient Routed Fine-Tuning for Mixture-of-Expert Model | Nov 12, 2024 | Arithmetic ReasoningMixture-of-Experts | —Unverified | 0 |
| Personalised Federated Learning: A Combinational Approach | Aug 22, 2021 | Federated LearningKnowledge Distillation | —Unverified | 0 |
| Pheromone-based Learning of Optimal Reasoning Paths | Jan 31, 2025 | ARCGSM8K | —Unverified | 0 |
| Physics-Guided Problem Decomposition for Scaling Deep Learning of High-dimensional Eigen-Solvers: The Case of Schrödinger's Equation | Feb 12, 2022 | Mixture-of-ExpertsProblem Decomposition | —Unverified | 0 |
| PICO: Secure Transformers via Robust Prompt Isolation and Cybersecurity Oversight | Apr 26, 2025 | Mixture-of-ExpertsPICO | —Unverified | 0 |
| Pipeline MoE: A Flexible MoE Implementation with Pipeline Parallelism | Apr 22, 2023 | AllMixture-of-Experts | —Unverified | 0 |
| Plasticity-Aware Mixture of Experts for Learning Under QoE Shifts in Adaptive Video Streaming | Apr 14, 2025 | Mixture-of-Experts | —Unverified | 0 |
| PLUME: Polyhedral Learning Using Mixture of Experts | Apr 22, 2019 | Generalization BoundsMixture-of-Experts | —Unverified | 0 |
| Pluralistic Image Completion with Probabilistic Mixture-of-Experts | May 18, 2022 | DiversityMixture-of-Experts | —Unverified | 0 |
| Pluralistic Salient Object Detection | Sep 4, 2024 | Mixture-of-ExpertsObject | —Unverified | 0 |
| P-Mamba: Marrying Perona Malik Diffusion with Mamba for Efficient Pediatric Echocardiographic Left Ventricular Segmentation | Feb 13, 2024 | MambaMixture-of-Experts | —Unverified | 0 |
| PMoL: Parameter Efficient MoE for Preference Mixing of LLM Alignment | Nov 2, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Point-MoE: Towards Cross-Domain Generalization in 3D Semantic Segmentation via Mixture-of-Experts | May 29, 2025 | 3D Semantic SegmentationDomain Generalization | —Unverified | 0 |
| Polynomial-Spline Neural Networks with Exact Integrals | Oct 26, 2021 | Mixture-of-Expertsregression | —Unverified | 0 |