| UOE: Unlearning One Expert Is Enough For Mixture-of-experts LLMS | Nov 27, 2024 | Large Language ModelMixture-of-Experts | —Unverified | 0 |
| Upcycling Instruction Tuning from Dense to Mixture-of-Experts via Parameter Merging | Oct 2, 2024 | DiversityMixture-of-Experts | —Unverified | 0 |
| Upcycling Large Language Models into Mixture of Experts | Oct 10, 2024 | Mixture-of-ExpertsMMLU | —Unverified | 0 |
| Using Deep Mixture-of-Experts to Detect Word Meaning Shift for TempoWiC | Nov 7, 2022 | Data AugmentationMixture-of-Experts | —Unverified | 0 |
| Utility-Driven Speculative Decoding for Mixture-of-Experts | Jun 17, 2025 | GPULarge Language Model | —Unverified | 0 |
| Vanilla Transformers are Transfer Capability Teachers | Mar 4, 2024 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| Variational Distillation of Diffusion Policies into Mixture of Experts | Jun 18, 2024 | DenoisingMixture-of-Experts | —Unverified | 0 |
| Variational Mixture of Gaussian Process Experts | Dec 1, 2008 | Gaussian ProcessesMixture-of-Experts | —Unverified | 0 |
| ViMoE: An Empirical Study of Designing Vision Mixture-of-Experts | Oct 21, 2024 | image-classificationImage Classification | —Unverified | 0 |
| Visual Saliency Prediction Using a Mixture of Deep Neural Networks | Feb 1, 2017 | Mixture-of-ExpertsSaliency Prediction | —Unverified | 0 |