SOTAVerified

Mixture-of-Experts

Papers

Showing 2130 of 1312 papers

TitleStatusHype
DeepSeekMoE: Towards Ultimate Expert Specialization in Mixture-of-Experts Language ModelsCode5
Aria: An Open Multimodal Native Mixture-of-Experts ModelCode5
Parrot: Multilingual Visual Instruction TuningCode5
LLaMA-MoE: Building Mixture-of-Experts from LLaMA with Continual Pre-trainingCode5
Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by TencentCode5
Comet: Fine-grained Computation-communication Overlapping for Mixture-of-ExpertsCode5
OpenMoE: An Early Effort on Open Mixture-of-Experts Language ModelsCode5
Chatlaw: A Multi-Agent Collaborative Legal Assistant with Knowledge Graph Enhanced Mixture-of-Experts Large Language ModelCode5
Kimi-VL Technical ReportCode5
JetMoE: Reaching Llama2 Performance with 0.1M DollarsCode4
Show:102550
← PrevPage 3 of 132Next →

No leaderboard results yet.