SOTAVerified

Mixture-of-Experts

Papers

Showing 76100 of 1312 papers

TitleStatusHype
Integrating Dynamical Systems Learning with Foundational Models: A Meta-Evolutionary AI Framework for Clinical Trials0
RankLLM: A Python Package for Reranking with LLMsCode0
I2MoE: Interpretable Multimodal Interaction-aware Mixture-of-ExpertsCode2
On Minimax Estimation of Parameters in Softmax-Contaminated Mixture of Experts0
TrajMoE: Spatially-Aware Mixture of Experts for Unified Human Mobility Modeling0
ThanoRA: Task Heterogeneity-Aware Multi-Task Low-Rank AdaptationCode1
μ-MoE: Test-Time Pruning as Micro-Grained Mixture-of-Experts0
Mod-Adapter: Tuning-Free and Versatile Multi-concept Personalization via Modulation Adapter0
Guiding the Experts: Semantic Priors for Efficient and Focused MoE RoutingCode0
EvidenceMoE: A Physics-Guided Mixture-of-Experts with Evidential Critics for Advancing Fluorescence Light Detection and Ranging in Scattering Media0
JanusDNA: A Powerful Bi-directional Hybrid DNA Foundation ModelCode1
DualComp: End-to-End Learning of a Unified Dual-Modality Lossless Compressor0
DriveMoE: Mixture-of-Experts for Vision-Language-Action Model in End-to-End Autonomous Driving0
Not All Models Suit Expert Offloading: On Local Routing Consistency of Mixture-of-Expert ModelsCode0
CoLA: Collaborative Low-Rank AdaptationCode0
Time Tracker: Mixture-of-Experts-Enhanced Foundation Time Series Forecasting Model with Decoupled Training Pipelines0
MoRE-Brain: Routed Mixture of Experts for Interpretable and Generalizable Cross-Subject fMRI Visual DecodingCode0
Hunyuan-TurboS: Advancing Large Language Models through Mamba-Transformer Synergy and Adaptive Chain-of-Thought0
Efficient Data Driven Mixture-of-Expert Extraction from Trained Networks0
Multimodal Cultural Safety: Evaluation Frameworks and Alignment StrategiesCode0
Balanced and Elastic End-to-end Training of Dynamic LLMs0
Multimodal Mixture of Low-Rank Experts for Sentiment Analysis and Emotion Recognition0
THOR-MoE: Hierarchical Task-Guided and Context-Responsive Routing for Neural Machine Translation0
Two Experts Are All You Need for Steering Thinking: Reinforcing Cognitive Effort in MoE Reasoning Models Without Additional Training0
FuxiMT: Sparsifying Large Language Models for Chinese-Centric Multilingual Machine Translation0
Show:102550
← PrevPage 4 of 53Next →

No leaderboard results yet.