SOTAVerified

Mixture-of-Experts

Papers

Showing 431440 of 1312 papers

TitleStatusHype
How Lightweight Can A Vision Transformer Be0
FedMerge: Federated Personalization via Model Merging0
Gating Dropout: Communication-efficient Regularization for Sparsely Activated Transformers0
A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts0
Coordination with Humans via Strategy Matching0
GEMNET: Effective Gated Gazetteer Representations for Recognizing Complex Entities in Low-context Input0
Generalizable Person Re-identification with Relevance-aware Mixture of Experts0
Generalization Error Analysis for Sparse Mixture-of-Experts: A Preliminary Study0
How to Upscale Neural Networks with Scaling Law? A Survey and Practical Guidelines0
Hunyuan-TurboS: Advancing Large Language Models through Mamba-Transformer Synergy and Adaptive Chain-of-Thought0
Show:102550
← PrevPage 44 of 132Next →

No leaderboard results yet.