SOTAVerified

Mixture-of-Experts

Papers

Showing 731740 of 1312 papers

TitleStatusHype
MoIN: Mixture of Introvert Experts to Upcycle an LLM0
GETS: Ensemble Temperature Scaling for Calibration in Graph Neural Networks0
AT-MoE: Adaptive Task-planning Mixture of Experts via LoRA Approach0
Upcycling Large Language Models into Mixture of Experts0
More Experts Than Galaxies: Conditionally-overlapping Experts With Biologically-Inspired Fixed RoutingCode0
Mono-InternVL: Pushing the Boundaries of Monolithic Multimodal Large Language Models with Endogenous Visual Pre-training0
Functional-level Uncertainty Quantification for Calibrated Fine-tuning on LLMs0
Toward generalizable learning of all (linear) first-order methods via memory augmented Transformers0
Scaling Laws Across Model Architectures: A Comparative Analysis of Dense and MoE Models in Large Language Models0
Probing the Robustness of Theory of Mind in Large Language Models0
Show:102550
← PrevPage 74 of 132Next →

No leaderboard results yet.