SOTAVerified

Mixture-of-Experts

Papers

Showing 481490 of 1312 papers

TitleStatusHype
Enhancing Generalization in Sparse Mixture of Experts Models: The Case for Increased Expert Activation in Compositional Tasks0
On the Risk of Evidence Pollution for Malicious Social Text Detection in the Era of LLMs0
EPS-MoE: Expert Pipeline Scheduler for Cost-Efficient MoE Inference0
Understanding Expert Structures on Minimax Parameter Estimation in Contaminated Mixture of Experts0
MoE-Pruner: Pruning Mixture-of-Experts Large Language Model using the Hints from Its Router0
Transformer Layer Injection: A Novel Approach for Efficient Upscaling of Large Language Models0
Quadratic Gating Functions in Mixture of Experts: A Statistical Insight0
GaVaMoE: Gaussian-Variational Gated Mixture of Experts for Explainable RecommendationCode1
MoH: Multi-Head Attention as Mixture-of-Head AttentionCode4
AlphaLoRA: Assigning LoRA Experts Based on Layer Training QualityCode1
Show:102550
← PrevPage 49 of 132Next →

No leaderboard results yet.