SOTAVerified

Mixture-of-Experts

Papers

Showing 226250 of 1312 papers

TitleStatusHype
Exploring Sparse MoE in GANs for Text-conditioned Image SynthesisCode1
Exploiting Inter-Layer Expert Affinity for Accelerating Mixture-of-Experts Model InferenceCode1
Few-Shot and Continual Learning with Attentive Independent MechanismsCode1
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided AdaptationCode1
LIBMoE: A Library for comprehensive benchmarking Mixture of Experts in Large Language ModelsCode1
Multi-Task Reinforcement Learning with Mixture of Orthogonal ExpertsCode1
A Time Series is Worth Five Experts: Heterogeneous Mixture of Experts for Traffic Flow PredictionCode1
AlphaLoRA: Assigning LoRA Experts Based on Layer Training QualityCode1
Enhancing NeRF akin to Enhancing LLMs: Generalizable NeRF Transformer with Mixture-of-View-ExpertsCode1
Enhancing Fast Feed Forward Networks with Load Balancing and a Master Leaf NodeCode1
Mixture of Experts Meets Prompt-Based Continual LearningCode1
BrainMAP: Learning Multiple Activation Pathways in Brain NetworksCode1
Emotion-Qwen: Training Hybrid Experts for Unified Emotion and General Vision-Language UnderstandingCode1
XMoE: Sparse Models with Fine-grained and Adaptive Expert SelectionCode1
Deep learning techniques for blind image super-resolution: A high-scale multi-domain perspective evaluationCode1
Mixture-of-Linear-Experts for Long-term Time Series ForecastingCode1
Mixture of Attention Heads: Selecting Attention Heads Per TokenCode1
Mixture of Decision Trees for Interpretable Machine LearningCode1
Emergent Modularity in Pre-trained TransformersCode1
EWMoE: An effective model for global weather forecasting with mixture-of-expertsCode1
Mixture of Experts Made Personalized: Federated Prompt Learning for Vision-Language ModelsCode1
Efficient Expert Pruning for Sparse Mixture-of-Experts Language Models: Enhancing Performance and Reducing Inference CostsCode1
MixPHM: Redundancy-Aware Parameter-Efficient Tuning for Low-Resource Visual Question AnsweringCode1
Efficient Fine-tuning of Audio Spectrogram Transformers via Soft Mixture of AdaptersCode1
BiMediX: Bilingual Medical Mixture of Experts LLMCode1
Show:102550
← PrevPage 10 of 53Next →

No leaderboard results yet.