SOTAVerified

Mixture-of-Experts

Papers

Showing 851875 of 1312 papers

TitleStatusHype
Revolutionizing Disease Diagnosis with simultaneous functional PET/MR and Deeply Integrated Brain Metabolic, Hemodynamic, and Perfusion Networks0
Ring-lite: Scalable Reasoning via C3PO-Stabilized Reinforcement Learning for LLMs0
RingMoE: Mixture-of-Modality-Experts Multi-Modal Foundation Models for Universal Remote Sensing Image Interpretation0
Robust and Explainable Depression Identification from Speech Using Vowel-Based Ensemble Learning Approaches0
Robust Audiovisual Speech Recognition Models with Mixture-of-Experts0
Robust Calibration For Improved Weather Prediction Under Distributional Shift0
Robust mixture of experts modeling using the skew t distribution0
Robust mixture of experts modeling using the t distribution0
RocketPPA: Code-Level Power, Performance, and Area Prediction via LLM and Mixture of Experts0
Routers in Vision Mixture of Experts: An Empirical Study0
RS-MoE: Mixture of Experts for Remote Sensing Image Captioning and Visual Question Answering0
RTM Ensemble Learning Results at Quality Estimation Task0
RTM Stacking Results for Machine Translation Performance Prediction0
RTM Super Learner Results at Quality Estimation Task0
S2MoE: Robust Sparse Mixture of Experts via Stochastic Learning0
Safe Real-World Autonomous Driving by Learning to Predict and Plan with a Mixture of Experts0
SAFEx: Analyzing Vulnerabilities of MoE-Based LLMs via Stable Safety-critical Expert Identification0
SAM-Med3D-MoE: Towards a Non-Forgetting Segment Anything Model via Mixture of Experts for 3D Medical Image Segmentation0
Scalable and Efficient MoE Training for Multitask Multilingual Models0
Scalable Multi-Domain Adaptation of Language Models using Modular Experts0
Scalable Neural Data Server: A Data Recommender for Transfer Learning0
Scaling and Enhancing LLM-based AVSR: A Sparse Mixture of Projectors Approach0
Scaling Intelligence: Designing Data Centers for Next-Gen Language Models0
Scaling Laws Across Model Architectures: A Comparative Analysis of Dense and MoE Models in Large Language Models0
Scaling Laws for Native Multimodal Models Scaling Laws for Native Multimodal Models0
Show:102550
← PrevPage 35 of 53Next →

No leaderboard results yet.