SOTAVerified

Mixture-of-Experts

Papers

Showing 126150 of 1312 papers

TitleStatusHype
Seed1.5-VL Technical Report0
QoS-Efficient Serving of Multiple Mixture-of-Expert LLMs Using Partial Runtime Reconfiguration0
Emotion-Qwen: Training Hybrid Experts for Unified Emotion and General Vision-Language UnderstandingCode1
Gated Attention for Large Language Models: Non-linearity, Sparsity, and Attention-Sink-FreeCode4
FloE: On-the-Fly MoE Inference on Memory-constrained GPU0
MxMoE: Mixed-precision Quantization for MoE with Accuracy and Performance Co-DesignCode1
Divide-and-Conquer: Cold-Start Bundle Recommendation via Mixture of Diffusion Experts0
Pangu Ultra MoE: How to Train Your Big MoE on Ascend NPUs0
SToLa: Self-Adaptive Touch-Language Framework with Tactile Commonsense Reasoning in Open-Ended Scenarios0
LLM-e Guess: Can LLMs Capabilities Advance Without Hardware Progress?Code0
STAR-Rec: Making Peace with Length Variance and Pattern Diversity in Sequential Recommendation0
Faster MoE LLM Inference for Extremely Large Models0
3D Gaussian Splatting Data Compression with Mixture of Priors0
Towards Smart Point-and-Shoot Photography0
Multimodal Deep Learning-Empowered Beam Prediction in Future THz ISAC Systems0
Optimizing LLMs for Resource-Constrained Environments: A Survey of Model Compression Techniques0
Finger Pose Estimation for Under-screen Fingerprint SensorCode0
Learning Heterogeneous Mixture of Scene Experts for Large-scale Neural Radiance FieldsCode3
Perception-Informed Neural Networks: Beyond Physics-Informed Neural Networks0
CoCoAFusE: Beyond Mixtures of Experts via Model Fusion0
CICADA: Cross-Domain Interpretable Coding for Anomaly Detection and Adaptation in Multivariate Time Series0
Mixture of Sparse Attention: Content-Based Learnable Sparse Attention via Expert-Choice RoutingCode1
MoxE: Mixture of xLSTM Experts with Entropy-Aware Routing for Efficient Language Modeling0
MicarVLMoE: A Modern Gated Cross-Aligned Vision-Language Mixture of Experts Model for Medical Image Captioning and Report GenerationCode0
Accelerating Mixture-of-Experts Training with Adaptive Expert Replication0
Show:102550
← PrevPage 6 of 53Next →

No leaderboard results yet.