| PROPER: A Progressive Learning Framework for Personalized Large Language Models with Group-Level Adaptation | Mar 3, 2025 | Mixture-of-Expertsparameter-efficient fine-tuning | —Unverified | 0 |
| Unify and Anchor: A Context-Aware Transformer for Cross-Domain Time Series Forecasting | Mar 3, 2025 | Domain GeneralizationMixture-of-Experts | —Unverified | 0 |
| DeRS: Towards Extremely Efficient Upcycled Mixture-of-Experts Models | Mar 3, 2025 | Mixture-of-ExpertsQuantization | —Unverified | 0 |
| Explainable Classifier for Malignant Lymphoma Subtyping via Cell Graph and Image Fusion | Mar 2, 2025 | Mixture-of-Expertswhole slide images | —Unverified | 0 |
| CL-MoE: Enhancing Multimodal Large Language Model with Dual Momentum Mixture-of-Experts for Continual Visual Question Answering | Mar 1, 2025 | Continual LearningLanguage Modeling | —Unverified | 0 |
| CoSMoEs: Compact Sparse Mixture of Experts | Feb 28, 2025 | Mixture-of-Experts | —Unverified | 0 |
| UniCodec: Unified Audio Codec with Single Domain-Adaptive Codebook | Feb 27, 2025 | Language ModelingLanguage Modelling | —Unverified | 0 |
| Mixture of Experts for Recognizing Depression from Interview and Reading Tasks | Feb 27, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Mixture of Experts-augmented Deep Unfolding for Activity Detection in IRS-aided Systems | Feb 27, 2025 | Action DetectionActivity Detection | —Unverified | 0 |
| OneRec: Unifying Retrieve and Rank with Generative Recommender and Iterative Preference Alignment | Feb 26, 2025 | Mixture-of-ExpertsRecommendation Systems | —Unverified | 0 |
| Drop-Upcycling: Training Sparse Mixture of Experts with Partial Re-initialization | Feb 26, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Evaluating Expert Contributions in a MoE LLM for Quiz-Based Tasks | Feb 24, 2025 | Mixture-of-ExpertsMMLU | —Unverified | 0 |
| ENACT-Heart -- ENsemble-based Assessment Using CNN and Transformer on Heart Sounds | Feb 24, 2025 | DiagnosticMixture-of-Experts | —Unverified | 0 |
| The Empirical Impact of Reducing Symmetries on the Performance of Deep Ensembles and MoE | Feb 24, 2025 | Linear Mode ConnectivityMixture-of-Experts | —Unverified | 0 |
| BigMac: A Communication-Efficient Mixture-of-Experts Model Structure for Fast Training and Inference | Feb 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| An Autonomous Network Orchestration Framework Integrating Large Language Models with Continual Reinforcement Learning | Feb 22, 2025 | ARCContinual Learning | —Unverified | 0 |
| Tight Clusters Make Specialized Experts | Feb 21, 2025 | ClusteringLanguage Modeling | CodeCode Available | 0 |
| Binary-Integer-Programming Based Algorithm for Expert Load Balancing in Mixture-of-Experts Models | Feb 21, 2025 | Mixture-of-Experts | CodeCode Available | 0 |
| Ray-Tracing for Conditionally Activated Neural Networks | Feb 20, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Unraveling the Localized Latents: Learning Stratified Manifold Structures in LLM Embedding Space with Sparse Mixture-of-Experts | Feb 19, 2025 | Dictionary LearningMixture-of-Experts | —Unverified | 0 |
| Every Expert Matters: Towards Effective Knowledge Distillation for Mixture-of-Experts Language Models | Feb 18, 2025 | Knowledge DistillationMixture-of-Experts | —Unverified | 0 |
| DSMoE: Matrix-Partitioned Experts with Dynamic Routing for Computation-Efficient Dense LLMs | Feb 18, 2025 | Computational EfficiencyLanguage Modeling | —Unverified | 0 |
| How to Upscale Neural Networks with Scaling Law? A Survey and Practical Guidelines | Feb 17, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Connector-S: A Survey of Connectors in Multi-modal Large Language Models | Feb 17, 2025 | Mixture-of-ExpertsSurvey | —Unverified | 0 |
| Fate: Fast Edge Inference of Mixture-of-Experts Models via Cross-Layer Gate | Feb 17, 2025 | GPUMixture-of-Experts | CodeCode Available | 0 |