SOTAVerified

Mixture-of-Experts

Papers

Showing 526550 of 1312 papers

TitleStatusHype
Accelerating MoE Model Inference with Expert Sharding0
GM-MoE: Low-Light Enhancement with Gated-Mechanism Mixture-of-Experts0
eMoE: Task-aware Memory Efficient Mixture-of-Experts-Based (MoE) Model Inference0
ResMoE: Space-efficient Compression of Mixture of Experts LLMs via Residual RestorationCode0
MoFE: Mixture of Frozen Experts Architecture0
Swift Hydra: Self-Reinforcing Generative Framework for Anomaly Detection with Multiple Mamba ModelsCode0
MANDARIN: Mixture-of-Experts Framework for Dynamic Delirium and Coma Prediction in ICU Patients: Development and Validation of an Acute Brain Dysfunction Prediction Model0
A Novel Trustworthy Video Summarization Algorithm Through a Mixture of LoRA Experts0
MoEMoE: Question Guided Dense and Scalable Sparse Mixture-of-Expert for Multi-source Multi-modal Answering0
FMT:A Multimodal Pneumonia Detection Model Based on Stacking MOE Framework0
Every FLOP Counts: Scaling a 300B Mixture-of-Experts LING LLM without Premium GPUs0
Capacity-Aware Inference: Mitigating the Straggler Effect in Mixture of Experts0
Symbolic Mixture-of-Experts: Adaptive Skill-based Routing for Heterogeneous Reasoning0
TS-RAG: Retrieval-Augmented Generation based Time Series Foundation Models are Stronger Zero-Shot Forecaster0
Continual Pre-training of MoEs: How robust is your router?0
A Generalist Cross-Domain Molecular Learning Framework for Structure-Based Drug Discovery0
Predictable Scale: Part I -- Optimal Hyperparameter Scaling Law in Large Language Model Pretraining0
Speculative MoE: Communication Efficient Parallel MoE Inference with Speculative Token and Expert Pre-scheduling0
Convergence Rates for Softmax Gating Mixture of Experts0
BrainNet-MoE: Brain-Inspired Mixture-of-Experts Learning for Neurological Disease Identification0
VoiceGRPO: Modern MoE Transformers with Group Relative Policy Optimization GRPO for AI Voice Health Care Applications on Voice Pathology DetectionCode0
Tabby: Tabular Data Synthesis with Language Models0
Union of Experts: Adapting Hierarchical Routing to Equivalently Decomposed TransformerCode0
How Do Consumers Really Choose: Exposing Hidden Preferences with the Mixture of Experts Model0
PROPER: A Progressive Learning Framework for Personalized Large Language Models with Group-Level Adaptation0
Show:102550
← PrevPage 22 of 53Next →

No leaderboard results yet.