| Multi-Treatment Multi-Task Uplift Modeling for Enhancing User Growth | Aug 23, 2024 | Causal InferenceMixture-of-Experts | —Unverified | 0 |
| DutyTTE: Deciphering Uncertainty in Origin-Destination Travel Time Estimation | Aug 23, 2024 | Deep Reinforcement LearningMixture-of-Experts | CodeCode Available | 0 |
| SQL-GEN: Bridging the Dialect Gap for Text-to-SQL Via Synthetic Data And Model Merging | Aug 22, 2024 | DiversityMixture-of-Experts | —Unverified | 0 |
| Jamba-1.5: Hybrid Transformer-Mamba Models at Scale | Aug 22, 2024 | ChatbotInstruction Following | CodeCode Available | 5 |
| Improving Factuality in Large Language Models via Decoding-Time Hallucinatory and Truthful Comparators | Aug 22, 2024 | HallucinationMixture-of-Experts | CodeCode Available | 0 |
| MoE-LPR: Multilingual Extension of Large Language Models through Mixture-of-Experts with Language Priors Routing | Aug 21, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| FedMoE: Personalized Federated Learning via Heterogeneous Mixture of Experts | Aug 21, 2024 | Federated LearningHeuristic Search | —Unverified | 0 |
| KAN4TSF: Are KAN and KAN-based models Effective for Time Series Forecasting? | Aug 21, 2024 | Mixture-of-ExpertsTime Series | CodeCode Available | 2 |
| HMoE: Heterogeneous Mixture of Experts for Language Modeling | Aug 20, 2024 | Computational EfficiencyLanguage Modeling | —Unverified | 0 |
| Navigating Spatio-Temporal Heterogeneity: A Graph Transformer Approach for Traffic Forecasting | Aug 20, 2024 | AttributeMixture-of-Experts | CodeCode Available | 1 |
| AnyGraph: Graph Foundation Model in the Wild | Aug 20, 2024 | Graph LearningMixture-of-Experts | CodeCode Available | 3 |
| AdapMoE: Adaptive Sensitivity-based Expert Gating and Management for Efficient MoE Inference | Aug 19, 2024 | ManagementMixture-of-Experts | CodeCode Available | 1 |
| A Unified Framework for Iris Anti-Spoofing: Introducing IrisGeneral Dataset and Masked-MoE Method | Aug 19, 2024 | Iris RecognitionMixture-of-Experts | —Unverified | 0 |
| Customizing Language Models with Instance-wise LoRA for Sequential Recommendation | Aug 19, 2024 | Mixture-of-ExpertsMulti-Task Learning | CodeCode Available | 1 |
| FEDKIM: Adaptive Federated Knowledge Injection into Medical Foundation Models | Aug 17, 2024 | Federated LearningMixture-of-Experts | CodeCode Available | 0 |
| Integrating Multi-view Analysis: Multi-view Mixture-of-Expert for Textual Personality Detection | Aug 16, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language Models | Aug 15, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| BAM! Just Like That: Simple and Efficient Parameter Upcycling for Mixture of Experts | Aug 15, 2024 | Mixture-of-Experts | —Unverified | 0 |
| A Survey on Model MoErging: Recycling and Routing Among Specialized Experts for Collaborative Learning | Aug 13, 2024 | Mixture-of-ExpertsSurvey | —Unverified | 0 |
| AquilaMoE: Efficient Training for MoE Models with Scale-Up and Scale-Out Strategies | Aug 13, 2024 | Language ModellingMixture-of-Experts | CodeCode Available | 1 |
| Layerwise Recurrent Router for Mixture-of-Experts | Aug 13, 2024 | AttributeMixture-of-Experts | CodeCode Available | 1 |
| HoME: Hierarchy of Multi-Gate Experts for Multi-Task Learning at Kuaishou | Aug 10, 2024 | Mixture-of-ExpertsMulti-Task Learning | —Unverified | 0 |
| Understanding the Performance and Estimating the Cost of LLM Fine-Tuning | Aug 8, 2024 | GPUMixture-of-Experts | CodeCode Available | 0 |
| LaDiMo: Layer-wise Distillation Inspired MoEfier | Aug 8, 2024 | Knowledge DistillationMixture-of-Experts | —Unverified | 0 |
| MoC-System: Efficient Fault Tolerance for Sparse Mixture-of-Experts Model Training | Aug 8, 2024 | Mixture-of-Experts | —Unverified | 0 |