SOTAVerified

Mixture-of-Experts

Papers

Showing 111120 of 1312 papers

TitleStatusHype
LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-TrainingCode2
LoRA-IR: Taming Low-Rank Experts for Efficient All-in-One Image RestorationCode2
Linear-MoE: Linear Sequence Modeling Meets Mixture-of-ExpertsCode2
Make LoRA Great Again: Boosting LoRA with Adaptive Singular Values and Mixture-of-Experts Optimization AlignmentCode2
Learning Robust Stereo Matching in the Wild with Selective Mixture-of-ExpertsCode2
SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language ModelsCode2
Aurora:Activating Chinese chat capability for Mixtral-8x7B sparse Mixture-of-Experts through Instruction-TuningCode2
Learning A Sparse Transformer Network for Effective Image DerainingCode2
KAN4TSF: Are KAN and KAN-based models Effective for Time Series Forecasting?Code2
LiMoE: Mixture of LiDAR Representation Learners from Automotive ScenesCode2
Show:102550
← PrevPage 12 of 132Next →

No leaderboard results yet.