| PICO: Secure Transformers via Robust Prompt Isolation and Cybersecurity Oversight | Apr 26, 2025 | Mixture-of-ExpertsPICO | —Unverified | 0 |
| NoEsis: Differentially Private Knowledge Transfer in Modular LLM Adaptation | Apr 25, 2025 | Code CompletionMixture-of-Experts | —Unverified | 0 |
| Unveiling the Hidden: Movie Genre and User Bias in Spoiler Detection | Apr 24, 2025 | Graph AttentionMixture-of-Experts | CodeCode Available | 0 |
| BadMoE: Backdooring Mixture-of-Experts LLMs via Optimizing Routing Triggers and Infecting Dormant Experts | Apr 24, 2025 | Backdoor AttackMixture-of-Experts | —Unverified | 0 |
| Manifold Induced Biases for Zero-shot and Few-shot Detection of Generated Images | Apr 21, 2025 | Mixture-of-Experts | CodeCode Available | 1 |
| MoE Parallel Folding: Heterogeneous Parallelism Mappings for Efficient Large-Scale MoE Model Training with Megatron Core | Apr 21, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Distribution-aware Forgetting Compensation for Exemplar-Free Lifelong Person Re-identification | Apr 21, 2025 | Exemplar-FreeKnowledge Distillation | CodeCode Available | 1 |
| HAECcity: Open-Vocabulary Scene Understanding of City-Scale Point Clouds with Superpoint Graph Clustering | Apr 18, 2025 | ClusteringGraph Clustering | —Unverified | 0 |
| Multi-Type Context-Aware Conversational Recommender Systems via Mixture-of-Experts | Apr 18, 2025 | Mixture-of-ExpertsRecommendation Systems | —Unverified | 0 |
| D^2MoE: Dual Routing and Dynamic Scheduling for Efficient On-Device MoE-based LLM Serving | Apr 17, 2025 | Mixture-of-ExpertsModel Compression | —Unverified | 0 |
| Unveiling Hidden Collaboration within Mixture-of-Experts in Large Language Models | Apr 16, 2025 | Dictionary LearningMixture-of-Experts | —Unverified | 0 |
| Dense Backpropagation Improves Training for Sparse Mixture-of-Experts | Apr 16, 2025 | Mixture-of-Experts | CodeCode Available | 1 |
| Trend Filtered Mixture of Experts for Automated Gating of High-Frequency Flow Cytometry Data | Apr 16, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Plasticity-Aware Mixture of Experts for Learning Under QoE Shifts in Adaptive Video Streaming | Apr 14, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Mixture-of-Shape-Experts (MoSE): End-to-End Shape Dictionary Framework to Prompt SAM for Generalizable Medical Segmentation | Apr 13, 2025 | Dictionary LearningDomain Generalization | —Unverified | 0 |
| MoE-Lens: Towards the Hardware Limit of High-Throughput MoE LLM Serving Under Resource Constraints | Apr 12, 2025 | CPUGPU | —Unverified | 0 |
| RouterKT: Mixture-of-Experts for Knowledge Tracing | Apr 11, 2025 | Knowledge TracingMixture-of-Experts | CodeCode Available | 0 |
| Regularized infill criteria for multi-objective Bayesian optimization with application to aircraft design | Apr 11, 2025 | Bayesian Optimizationglobal-optimization | —Unverified | 0 |
| Seed1.5-Thinking: Advancing Superb Reasoning Models with Reinforcement Learning | Apr 10, 2025 | Mixture-of-Expertsreinforcement-learning | —Unverified | 0 |
| Kimi-VL Technical Report | Apr 10, 2025 | Long-Context UnderstandingMathematical Reasoning | CodeCode Available | 5 |
| C3PO: Critical-Layer, Core-Expert, Collaborative Pathway Optimization for Test-Time Expert Re-Mixing | Apr 10, 2025 | In-Context LearningMixture-of-Experts | CodeCode Available | 1 |
| Cluster-Driven Expert Pruning for Mixture-of-Experts Large Language Models | Apr 10, 2025 | Computational EfficiencyMixture-of-Experts | CodeCode Available | 0 |
| Scaling Laws for Native Multimodal Models Scaling Laws for Native Multimodal Models | Apr 10, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Adaptive Detection of Fast Moving Celestial Objects Using a Mixture of Experts and Physical-Inspired Neural Network | Apr 10, 2025 | Mixture-of-Expertsobject-detection | —Unverified | 0 |
| Holistic Capability Preservation: Towards Compact Yet Comprehensive Reasoning Models | Apr 9, 2025 | Instruction FollowingMathematical Problem-Solving | —Unverified | 0 |