SOTAVerified

Mixture-of-Experts

Papers

Showing 176200 of 1312 papers

TitleStatusHype
MiLo: Efficient Quantized MoE Inference with Mixture of Low-Rank CompensatorsCode1
Exploring Sparse MoE in GANs for Text-conditioned Image SynthesisCode1
Merge, Then Compress: Demystify Efficient SMoE with Hints from Its Routing PolicyCode1
PAD-Net: An Efficient Framework for Dynamic NetworksCode1
MEFT: Memory-Efficient Fine-Tuning through Sparse AdapterCode1
Merging Experts into One: Improving Computational Efficiency of Mixture of ExpertsCode1
Exploiting Inter-Layer Expert Affinity for Accelerating Mixture-of-Experts Model InferenceCode1
ChatVLA: Unified Multimodal Understanding and Robot Control with Vision-Language-Action ModelCode1
Examining Post-Training Quantization for Mixture-of-Experts: A BenchmarkCode1
MedCoT: Medical Chain of Thought via Hierarchical ExpertCode1
Meta-DMoE: Adapting to Domain Shift by Meta-Distillation from Mixture-of-ExpertsCode1
Merging Multi-Task Models via Weight-Ensembling Mixture of ExpertsCode1
Few-Shot and Continual Learning with Attentive Independent MechanismsCode1
Enhancing Fast Feed Forward Networks with Load Balancing and a Master Leaf NodeCode1
FineMoGen: Fine-Grained Spatio-Temporal Motion Generation and EditingCode1
Enhancing NeRF akin to Enhancing LLMs: Generalizable NeRF Transformer with Mixture-of-View-ExpertsCode1
Manifold Induced Biases for Zero-shot and Few-shot Detection of Generated ImagesCode1
M^4oE: A Foundation Model for Medical Multimodal Image Segmentation with Mixture of ExpertsCode1
Making Neural Networks Interpretable with Attribution: Application to Implicit Signals PredictionCode1
Mask and Reason: Pre-Training Knowledge Graph Transformers for Complex Logical QueriesCode1
Emergent Modularity in Pre-trained TransformersCode1
Emotion-Qwen: Training Hybrid Experts for Unified Emotion and General Vision-Language UnderstandingCode1
GaVaMoE: Gaussian-Variational Gated Mixture of Experts for Explainable RecommendationCode1
M^3ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task Learning with Model-Accelerator Co-designCode1
Addressing Confounding Feature Issue for Causal RecommendationCode1
Show:102550
← PrevPage 8 of 53Next →

No leaderboard results yet.