| BIG-MoE: Bypass Isolated Gating MoE for Generalized Multimodal Face Anti-Spoofing | Dec 24, 2024 | Decision MakingFace Anti-Spoofing | CodeCode Available | 0 |
| Fast filtering of non-Gaussian models using Amortized Optimal Transport Maps | Mar 16, 2025 | Mixture-of-Experts | CodeCode Available | 0 |
| A Gated Residual Kolmogorov-Arnold Networks for Mixtures of Experts | Sep 23, 2024 | Kolmogorov-Arnold NetworksMixture-of-Experts | CodeCode Available | 0 |
| Bidirectional Attention as a Mixture of Continuous Word Experts | Jul 8, 2023 | Language ModellingMixture-of-Experts | CodeCode Available | 0 |
| Universal Simultaneous Machine Translation with Mixture-of-Experts Wait-k Policy | Sep 11, 2021 | Machine TranslationMixture-of-Experts | CodeCode Available | 0 |
| Tight Clusters Make Specialized Experts | Feb 21, 2025 | ClusteringLanguage Modeling | CodeCode Available | 0 |
| CompeteSMoE -- Statistically Guaranteed Mixture of Experts Training via Competition | May 19, 2025 | Mixture-of-Experts | CodeCode Available | 0 |
| Two Heads are Better than One: Nested PoE for Robust Defense Against Multi-Backdoors | Apr 2, 2024 | Data PoisoningHate Speech Detection | CodeCode Available | 0 |
| LLM-e Guess: Can LLMs Capabilities Advance Without Hardware Progress? | May 7, 2025 | Large Language ModelMixture-of-Experts | CodeCode Available | 0 |
| FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models | Aug 15, 2024 | Mixture-of-Experts | CodeCode Available | 0 |