SOTAVerified

Mixture-of-Experts

Papers

Showing 201250 of 1312 papers

TitleStatusHype
Gradient-free variational learning with conditional mixture networksCode1
Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer PruningCode1
Gated Multimodal Units for Information FusionCode1
PAD-Net: An Efficient Framework for Dynamic NetworksCode1
HyperRouter: Towards Efficient Training and Inference of Sparse Mixture of ExpertsCode1
HyperMoE: Towards Better Mixture of Experts via Transferring Among ExpertsCode1
GaVaMoE: Gaussian-Variational Gated Mixture of Experts for Explainable RecommendationCode1
MoEDiff-SR: Mixture of Experts-Guided Diffusion Model for Region-Adaptive MRI Super-ResolutionCode1
Heterogeneous Mixture of Experts for Remote Sensing Image Super-ResolutionCode1
ChatVLA: Unified Multimodal Understanding and Robot Control with Vision-Language-Action ModelCode1
FLAME-MoE: A Transparent End-to-End Research Platform for Mixture-of-Experts Language ModelsCode1
Model-GLUE: Democratized LLM Scaling for A Large Model Zoo in the WildCode1
MoCaE: Mixture of Calibrated Experts Significantly Improves Object DetectionCode1
Contrastive Learning and Mixture of Experts Enables Precise Vector EmbeddingsCode1
Modality Interactive Mixture-of-Experts for Fake News DetectionCode1
Few-Shot and Continual Learning with Attentive Independent MechanismsCode1
FineMoGen: Fine-Grained Spatio-Temporal Motion Generation and EditingCode1
Addressing Confounding Feature Issue for Causal RecommendationCode1
C3PO: Critical-Layer, Core-Expert, Collaborative Pathway Optimization for Test-Time Expert Re-MixingCode1
FreqMoE: Enhancing Time Series Forecasting through Frequency Decomposition Mixture of ExpertsCode1
Mixture of Sparse Attention: Content-Based Learnable Sparse Attention via Expert-Choice RoutingCode1
Specialized federated learning using a mixture of expertsCode1
Mixture-of-Linear-Experts for Long-term Time Series ForecastingCode1
Heterogeneous Multi-task Learning with Expert DiversityCode1
Exploring Sparse MoE in GANs for Text-conditioned Image SynthesisCode1
Learning Soccer Juggling Skills with Layer-wise Mixture-of-ExpertsCode1
Mixture of Experts Made Personalized: Federated Prompt Learning for Vision-Language ModelsCode1
Exploiting Inter-Layer Expert Affinity for Accelerating Mixture-of-Experts Model InferenceCode1
Examining Post-Training Quantization for Mixture-of-Experts: A BenchmarkCode1
Frequency-Adaptive Pan-Sharpening with Mixture of ExpertsCode1
DAMEX: Dataset-aware Mixture-of-Experts for visual understanding of mixture-of-datasetsCode1
A Time Series is Worth Five Experts: Heterogeneous Mixture of Experts for Traffic Flow PredictionCode1
Mixture of Experts Meets Prompt-Based Continual LearningCode1
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided AdaptationCode1
Enhancing Fast Feed Forward Networks with Load Balancing and a Master Leaf NodeCode1
XMoE: Sparse Models with Fine-grained and Adaptive Expert SelectionCode1
Enhancing NeRF akin to Enhancing LLMs: Generalizable NeRF Transformer with Mixture-of-View-ExpertsCode1
Lifting the Curse of Capacity Gap in Distilling Language ModelsCode1
BrainMAP: Learning Multiple Activation Pathways in Brain NetworksCode1
Emergent Modularity in Pre-trained TransformersCode1
Emotion-Qwen: Training Hybrid Experts for Unified Emotion and General Vision-Language UnderstandingCode1
MixPHM: Redundancy-Aware Parameter-Efficient Tuning for Low-Resource Visual Question AnsweringCode1
Efficient Dictionary Learning with Switch Sparse AutoencodersCode1
MiCE: Mixture of Contrastive Experts for Unsupervised Image ClusteringCode1
Efficient Expert Pruning for Sparse Mixture-of-Experts Language Models: Enhancing Performance and Reducing Inference CostsCode1
BiMediX: Bilingual Medical Mixture of Experts LLMCode1
MeteoRA: Multiple-tasks Embedded LoRA for Large Language ModelsCode1
MiLo: Efficient Quantized MoE Inference with Mixture of Low-Rank CompensatorsCode1
Merging Multi-Task Models via Weight-Ensembling Mixture of ExpertsCode1
Efficient and Degradation-Adaptive Network for Real-World Image Super-ResolutionCode1
Show:102550
← PrevPage 5 of 27Next →

No leaderboard results yet.