| Sigmoid Self-Attention has Lower Sample Complexity than Softmax Self-Attention: A Mixture-of-Experts Perspective | Feb 1, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Adaptive Prompt: Unlocking the Power of Visual Prompt Tuning | Jan 31, 2025 | Mixture-of-ExpertsVisual Prompt Tuning | —Unverified | 0 |
| Pheromone-based Learning of Optimal Reasoning Paths | Jan 31, 2025 | ARCGSM8K | —Unverified | 0 |
| MolGraph-xLSTM: A graph-based dual-level xLSTM framework with multi-head mixture-of-experts for enhanced molecular representation and interpretability | Jan 30, 2025 | Drug DiscoveryMixture-of-Experts | —Unverified | 0 |
| Heuristic-Informed Mixture of Experts for Link Prediction in Multilayer Networks | Jan 29, 2025 | Link PredictionMixture-of-Experts | —Unverified | 0 |
| Free Agent in Agent-Based Mixture-of-Experts Generative AI Framework | Jan 29, 2025 | Fraud DetectionMixture-of-Experts | —Unverified | 0 |
| 3D-MoE: A Mixture-of-Experts Multi-modal LLM for 3D Vision and Pose Diffusion via Rectified Flow | Jan 28, 2025 | Instruction FollowingMixture-of-Experts | —Unverified | 0 |
| Static Batching of Irregular Workloads on GPUs: Framework and Application to Efficient MoE Model Inference | Jan 27, 2025 | GPUMixture-of-Experts | —Unverified | 0 |
| ToMoE: Converting Dense Large Language Models to Mixture-of-Experts through Dynamic Structural Pruning | Jan 25, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Each Rank Could be an Expert: Single-Ranked Mixture of Experts LoRA for Multi-Task Learning | Jan 25, 2025 | Mixture-of-ExpertsMulti-Task Learning | —Unverified | 0 |
| Mean-field limit from general mixtures of experts to quantum neural networks | Jan 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Sparse Mixture-of-Experts for Non-Uniform Noise Reduction in MRI Images | Jan 24, 2025 | DenoisingDiagnostic | —Unverified | 0 |
| CSAOT: Cooperative Multi-Agent System for Active Object Tracking | Jan 23, 2025 | Autonomous NavigationDeep Reinforcement Learning | —Unverified | 0 |
| UniUIR: Considering Underwater Image Restoration as An All-in-One Learner | Jan 22, 2025 | AllDepth Estimation | —Unverified | 0 |
| BLR-MoE: Boosted Language-Routing Mixture of Experts for Domain-Robust Multilingual E2E ASR | Jan 22, 2025 | Mixture-of-Experts | —Unverified | 0 |
| LLM4WM: Adapting LLM for Wireless Multi-Tasking | Jan 22, 2025 | General KnowledgeLanguage Modeling | —Unverified | 0 |
| Autonomy-of-Experts Models | Jan 22, 2025 | Decision MakingMixture-of-Experts | —Unverified | 0 |
| Demons in the Detail: On Implementing Load Balancing Loss for Training Specialized Mixture-of-Expert Models | Jan 21, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Parameters vs FLOPs: Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models | Jan 21, 2025 | Mixture-of-Experts | —Unverified | 0 |
| SCFCRC: Simultaneously Counteract Feature Camouflage and Relation Camouflage for Fraud Detection | Jan 21, 2025 | Contrastive LearningFraud Detection | —Unverified | 0 |
| FSMoE: A Flexible and Scalable Training System for Sparse Mixture-of-Experts Models | Jan 18, 2025 | GPUMixture-of-Experts | —Unverified | 0 |
| OMoE: Diversifying Mixture of Low-Rank Adaptation by Orthogonal Finetuning | Jan 17, 2025 | Computational EfficiencyDiversity | —Unverified | 0 |
| LLM-Based Routing in Mixture of Experts: A Novel Framework for Trading | Jan 16, 2025 | Mixture-of-ExpertsWorld Knowledge | —Unverified | 0 |
| GRAPHMOE: Amplifying Cognitive Depth of Mixture-of-Experts Network via Introducing Self-Rethinking Mechanism | Jan 14, 2025 | Mixture-of-Experts | —Unverified | 0 |
| PSReg: Prior-guided Sparse Mixture of Experts for Point Cloud Registration | Jan 14, 2025 | Mixture-of-ExpertsPoint Cloud Registration | —Unverified | 0 |
| A Multi-Modal Deep Learning Framework for Pan-Cancer Prognosis | Jan 13, 2025 | Deep LearningMixture-of-Experts | CodeCode Available | 0 |
| TAMER: A Test-Time Adaptive MoE-Driven Framework for EHR Representation Learning | Jan 10, 2025 | Mixture-of-ExpertsRepresentation Learning | CodeCode Available | 0 |
| Optimizing Distributed Deployment of Mixture-of-Experts Model Inference in Serverless Computing | Jan 9, 2025 | Bayesian OptimizationCPU | —Unverified | 0 |
| mFabric: An Efficient and Scalable Fabric for Mixture-of-Experts Training | Jan 7, 2025 | BlockingGPU | —Unverified | 0 |
| Mixture-of-Experts Graph Transformers for Interpretable Particle Collision Detection | Jan 6, 2025 | Decision MakingMixture-of-Experts | CodeCode Available | 0 |
| Fresh-CL: Feature Realignment through Experts on Hypersphere in Continual Learning | Jan 4, 2025 | Continual LearningMixture-of-Experts | —Unverified | 0 |
| MoVE-KD: Knowledge Distillation for VLMs with Mixture of Visual Encoders | Jan 3, 2025 | Knowledge DistillationMixture-of-Experts | —Unverified | 0 |
| MExD: An Expert-Infused Diffusion Model for Whole-Slide Image Classification | Jan 1, 2025 | image-classificationImage Classification | —Unverified | 0 |
| Correlative and Discriminative Label Grouping for Multi-Label Visual Prompt Tuning | Jan 1, 2025 | image-classificationImage Classification | —Unverified | 0 |
| Towards Efficient Foundation Model for Zero-shot Amodal Segmentation | Jan 1, 2025 | Mixture-of-Experts | —Unverified | 0 |
| REM: A Scalable Reinforced Multi-Expert Framework for Multiplex Influence Maximization | Jan 1, 2025 | Mixture-of-Experts | —Unverified | 0 |
| UNIALIGN: Scaling Multimodal Alignment within One Unified Model | Jan 1, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Learning Heterogeneous Tissues with Mixture of Experts for Gigapixel Whole Slide Images | Jan 1, 2025 | Mixture-of-Expertswhole slide images | —Unverified | 0 |
| CNC: Cross-modal Normality Constraint for Unsupervised Multi-class Anomaly Detection | Dec 31, 2024 | Anomaly DetectionAttribute | —Unverified | 0 |
| Multimodal Variational Autoencoder: a Barycentric View | Dec 29, 2024 | Mixture-of-ExpertsRepresentation Learning | —Unverified | 0 |
| UniRestorer: Universal Image Restoration via Adaptively Estimating Image Degradation at Proper Granularity | Dec 28, 2024 | Image RestorationMixture-of-Experts | CodeCode Available | 0 |
| Graph Mixture of Experts and Memory-augmented Routers for Multivariate Time Series Anomaly Detection | Dec 26, 2024 | Anomaly DetectionMixture-of-Experts | —Unverified | 0 |
| AskChart: Universal Chart Understanding through Textual Enhancement | Dec 26, 2024 | Chart UnderstandingMixture-of-Experts | CodeCode Available | 0 |
| BIG-MoE: Bypass Isolated Gating MoE for Generalized Multimodal Face Anti-Spoofing | Dec 24, 2024 | Decision MakingFace Anti-Spoofing | CodeCode Available | 0 |
| UME: Upcycling Mixture-of-Experts for Scalable and Efficient Automatic Speech Recognition | Dec 23, 2024 | Automatic Speech RecognitionAutomatic Speech Recognition (ASR) | —Unverified | 0 |
| Part-Of-Speech Sensitivity of Routers in Mixture of Experts Models | Dec 22, 2024 | Mixture-of-ExpertsPOS | —Unverified | 0 |
| Theory of Mixture-of-Experts for Mobile Edge Computing | Dec 20, 2024 | Computational EfficiencyContinual Learning | —Unverified | 0 |
| SEKE: Specialised Experts for Keyword Extraction | Dec 18, 2024 | DescriptiveKeyword Extraction | CodeCode Available | 0 |
| SMOSE: Sparse Mixture of Shallow Experts for Interpretable Reinforcement Learning in Continuous Control Tasks | Dec 17, 2024 | continuous-controlContinuous Control | CodeCode Available | 0 |
| DAOP: Data-Aware Offloading and Predictive Pre-Calculation for Efficient MoE Inference | Dec 16, 2024 | CPUGPU | CodeCode Available | 0 |