| CoLA: Collaborative Low-Rank Adaptation | May 21, 2025 | CoLAMixture-of-Experts | CodeCode Available | 0 | 5 |
| Fast filtering of non-Gaussian models using Amortized Optimal Transport Maps | Mar 16, 2025 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| MoNTA: Accelerating Mixture-of-Experts Training with Network-Traffc-Aware Parallel Optimization | Nov 1, 2024 | 8kMixture-of-Experts | CodeCode Available | 0 | 5 |
| MoVEInt: Mixture of Variational Experts for Learning Human-Robot Interactions from Demonstrations | Jul 10, 2024 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| Nesti-Net: Normal Estimation for Unstructured 3D Point Clouds using Convolutional Neural Networks | Dec 3, 2018 | Mixture-of-ExpertsSurface Normals Estimation | CodeCode Available | 0 | 5 |
| FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models | Aug 15, 2024 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| Extreme Classification in Log Memory using Count-Min Sketch: A Case Study of Amazon Search with 50M Products | Oct 28, 2019 | ClassificationGeneral Classification | CodeCode Available | 0 | 5 |
| MoE-MLoRA for Multi-Domain CTR Prediction: Efficient Adaptation with Expert Specialization | Jun 9, 2025 | Click-Through Rate PredictionDiversity | CodeCode Available | 0 | 5 |
| MoE-LPR: Multilingual Extension of Large Language Models through Mixture-of-Experts with Language Priors Routing | Aug 21, 2024 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| Cluster-Driven Expert Pruning for Mixture-of-Experts Large Language Models | Apr 10, 2025 | Computational EfficiencyMixture-of-Experts | CodeCode Available | 0 | 5 |
| MoE-I^2: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank Decomposition | Nov 1, 2024 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| Exploring Model Consensus to Generate Translation Paraphrases | Jul 1, 2020 | DiversityMachine Translation | CodeCode Available | 0 | 5 |
| Exploiting Activation Sparsity with Dense to Dynamic-k Mixture-of-Experts Conversion | Oct 6, 2023 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| MoRE-Brain: Routed Mixture of Experts for Interpretable and Generalizable Cross-Subject fMRI Visual Decoding | May 21, 2025 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| More Experts Than Galaxies: Conditionally-overlapping Experts With Biologically-Inspired Fixed Routing | Oct 10, 2024 | image-classificationImage Classification | CodeCode Available | 0 | 5 |
| Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-Experts | Jul 19, 2018 | Binary ClassificationClick-Through Rate Prediction | CodeCode Available | 0 | 5 |
| MLP-KAN: Unifying Deep Representation and Function Learning | Oct 3, 2024 | Kolmogorov-Arnold NetworksMixture-of-Experts | CodeCode Available | 0 | 5 |
| Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-Experts | Jun 8, 2023 | Language ModelingLanguage Modelling | CodeCode Available | 0 | 5 |
| Expert Sample Consensus Applied to Camera Re-Localization | Aug 7, 2019 | Camera LocalizationMixture-of-Experts | CodeCode Available | 0 | 5 |
| Mixture of Modular Experts: Distilling Knowledge from a Multilingual Teacher into Specialized Modular Language Models | Jul 28, 2024 | Knowledge DistillationMixture-of-Experts | CodeCode Available | 0 | 5 |
| CompeteSMoE - Effective Training of Sparse Mixture of Experts via Competition | Feb 4, 2024 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| Mixture of Nested Experts: Adaptive Processing of Visual Tokens | Jul 29, 2024 | Mixture-of-Experts | CodeCode Available | 0 | 5 |
| Mixture of Link Predictors on Graphs | Feb 13, 2024 | Link PredictionMixture-of-Experts | CodeCode Available | 0 | 5 |
| Modality-Independent Brain Lesion Segmentation with Privacy-aware Continual Learning | Mar 26, 2025 | Continual LearningKnowledge Distillation | CodeCode Available | 0 | 5 |
| A non-asymptotic approach for model selection via penalization in high-dimensional mixture of experts models | Apr 6, 2021 | Mixture-of-ExpertsModel Selection | CodeCode Available | 0 | 5 |