| Omni-SMoLA: Boosting Generalist Multimodal Models with Soft Mixture of Low-rank Experts | Dec 1, 2023 | Chart Question AnsweringDocument AI | —Unverified | 0 |
| OMoE: Diversifying Mixture of Low-Rank Adaptation by Orthogonal Finetuning | Jan 17, 2025 | Computational EfficiencyDiversity | —Unverified | 0 |
| On component interactions in two-stage recommender systems | Jun 28, 2021 | Mixture-of-ExpertsRecommendation Systems | —Unverified | 0 |
| SkillNet-NLU: A Sparsely Activated Model for General-Purpose Natural Language Understanding | Mar 7, 2022 | Language ModellingMasked Language Modeling | —Unverified | 0 |
| OneRec: Unifying Retrieve and Rank with Generative Recommender and Iterative Preference Alignment | Feb 26, 2025 | Mixture-of-ExpertsRecommendation Systems | —Unverified | 0 |
| One Student Knows All Experts Know: From Sparse to Dense | Jan 26, 2022 | AllKnowledge Distillation | —Unverified | 0 |
| On Expert Estimation in Hierarchical Mixture of Experts: Beyond Softmax Gating Functions | Oct 3, 2024 | image-classificationImage Classification | —Unverified | 0 |
| On Least Square Estimation in Softmax Gating Mixture of Experts | Feb 5, 2024 | Mixture-of-Experts | —Unverified | 0 |
| On Minimax Estimation of Parameters in Softmax-Contaminated Mixture of Experts | May 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| On Multi-objective Policy Optimization as a Tool for Reinforcement Learning: Case Studies in Offline RL and Finetuning | Jun 15, 2021 | Deep Reinforcement LearningMixture-of-Experts | —Unverified | 0 |