| Interpretable Mixture of Experts | Jun 5, 2022 | Decision MakingMixture-of-Experts | —Unverified | 0 | 0 |
| Interpretable mixture of experts for time series prediction under recurrent and non-recurrent conditions | Sep 5, 2024 | Mixture-of-ExpertsTime Series | —Unverified | 0 | 0 |
| Intuition-aware Mixture-of-Rank-1-Experts for Parameter Efficient Finetuning | Apr 13, 2024 | DiversityMixture-of-Experts | —Unverified | 0 | 0 |
| Investigating Mixture of Experts in Dense Retrieval | Dec 16, 2024 | Information RetrievalMixture-of-Experts | —Unverified | 0 | 0 |
| Investigating the potential of Sparse Mixtures-of-Experts for multi-domain neural machine translation | Jul 1, 2024 | Machine TranslationMixture-of-Experts | —Unverified | 0 | 0 |
| Is Temperature Sample Efficient for Softmax Gaussian Mixture of Experts? | Jan 25, 2024 | Mixture-of-Expertsparameter estimation | —Unverified | 0 | 0 |
| JiuZhang 2.0: A Unified Chinese Pre-trained Language Model for Multi-task Mathematical Problem Solving | Jun 19, 2023 | In-Context LearningLanguage Modeling | —Unverified | 0 | 0 |
| Joint MoE Scaling Laws: Mixture of Experts Can Be Memory Efficient | Feb 7, 2025 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 | 0 |
| KAAE: Numerical Reasoning for Knowledge Graphs via Knowledge-aware Attributes Learning | Nov 20, 2024 | AttributeContrastive Learning | —Unverified | 0 | 0 |
| KAT-V1: Kwai-AutoThink Technical Report | Jul 11, 2025 | Knowledge DistillationLarge Language Model | —Unverified | 0 | 0 |