SOTAVerified

Mixture-of-Experts

Papers

Showing 931940 of 1312 papers

TitleStatusHype
Learning to Specialize: Joint Gating-Expert Training for Adaptive MoEs in Decentralized Settings0
ShiftAddViT: Mixture of Multiplication Primitives Towards Efficient Vision TransformerCode1
Attention Weighted Mixture of Experts with Contrastive Learning for Personalized Ranking in E-commerce0
Mixture-of-Supernets: Improving Weight-Sharing Supernet Training with Architecture-Routed Mixture-of-ExpertsCode0
ModuleFormer: Modularity Emerges from Mixture-of-ExpertsCode2
Patch-level Routing in Mixture-of-Experts is Provably Sample-efficient for Convolutional Neural NetworksCode1
COMET: Learning Cardinality Constrained Mixture of Experts with Trees and Local SearchCode1
Revisiting Hate Speech Benchmarks: From Data Curation to System DeploymentCode0
Divide, Conquer, and Combine: Mixture of Semantic-Independent Experts for Zero-Shot Dialogue State Tracking0
Edge-MoE: Memory-Efficient Multi-Task Vision Transformer Architecture with Task-level Sparsity via Mixture-of-ExpertsCode1
Show:102550
← PrevPage 94 of 132Next →

No leaderboard results yet.