SOTAVerified

Mixture-of-Experts

Papers

Showing 626650 of 1312 papers

TitleStatusHype
Connector-S: A Survey of Connectors in Multi-modal Large Language Models0
fMoE: Fine-Grained Expert Offloading for Large Mixture-of-Experts Serving0
FloE: On-the-Fly MoE Inference on Memory-constrained GPU0
Configurable Foundation Models: Building LLMs from a Modular Perspective0
FlexMoE: Scaling Large-scale Sparse Pre-trained Model Training via Dynamic Device Placement0
Mixture-of-Experts Meets Instruction Tuning:A Winning Combination for Large Language Models0
Conditional computation in neural networks: principles and research trends0
Fixing MoE Over-Fitting on Low-Resource Languages in Multilingual Machine Translation0
FinTeamExperts: Role Specialized MOEs For Financial Analysis0
On the Adaptation to Concept Drift for CTR Prediction0
A Review of Sparse Expert Models in Deep Learning0
FineQuant: Unlocking Efficiency with Fine-Grained Weight-Only Quantization for LLMs0
Finding Fantastic Experts in MoEs: A Unified Study for Expert Dropping Strategies and Observations0
Filtered not Mixed: Stochastic Filtering-Based Online Gating for Mixture of Large Language Models0
Complexity Experts are Task-Discriminative Learners for Any Image Restoration0
A Review of DeepSeek Models' Key Innovative Techniques0
AdaMV-MoE: Adaptive Multi-Task Vision Mixture-of-Experts0
Learning to Specialize: Joint Gating-Expert Training for Adaptive MoEs in Decentralized Settings0
FedMoE: Personalized Federated Learning via Heterogeneous Mixture of Experts0
FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation0
FedMerge: Federated Personalization via Model Merging0
A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts0
Affect in Tweets Using Experts Model0
Federated Mixture of Experts0
Federated learning using mixture of experts0
Show:102550
← PrevPage 26 of 53Next →

No leaderboard results yet.