SOTAVerified

Mixture-of-Experts

Papers

Showing 626650 of 1312 papers

TitleStatusHype
An Autonomous Network Orchestration Framework Integrating Large Language Models with Continual Reinforcement Learning0
AdaSpeech 3: Adaptive Text to Speech for Spontaneous Style0
MIRA: Medical Time Series Foundation Model for Real-World Health Data0
Lynx: Enabling Efficient MoE Inference through Dynamic Batch-Aware Expert Selection0
LSTM-based Mixture-of-Experts for Knowledge-Aware Dialogues0
Efficient Data Driven Mixture-of-Expert Extraction from Trained Networks0
M6-T: Exploring Sparse Expert Models and Beyond0
Machine learning based digital twin for dynamical systems with multiple time-scales0
LPT++: Efficient Training on Mixture of Long-tailed Experts0
Low-Rank Mixture-of-Experts for Continual Medical Image Segmentation0
MALoRA: Mixture of Asymmetric Low-Rank Adaptation for Enhanced Multi-Task Learning0
MANDARIN: Mixture-of-Experts Framework for Dynamic Delirium and Coma Prediction in ICU Patients: Development and Validation of an Acute Brain Dysfunction Prediction Model0
Boost Your NeRF: A Model-Agnostic Mixture of Experts Framework for High Quality and Efficient Rendering0
An Autonomous Negotiating Agent Framework with Reinforcement Learning Based Strategies and Adaptive Strategy Switching Mechanism0
Lory: Fully Differentiable Mixture-of-Experts for Autoregressive Language Model Pre-training0
LoRA-Switch: Boosting the Efficiency of Dynamic LLM Adapters via System-Algorithm Co-design0
Efficient Mixture-of-Expert for Video-based Driver State and Physiological Multi-task Estimation in Conditional Autonomous Driving0
Massively Multilingual Shallow Fusion with Large Language Models0
LoRA-Mixer: Coordinate Modular LoRA Experts Through Serial Attention Routing0
MAST-Pro: Dynamic Mixture-of-Experts for Adaptive Segmentation of Pan-Tumors with Knowledge-Driven Prompts0
Efficient and Effective Weight-Ensembling Mixture of Experts for Multi-Task Model Merging0
Adaptive Segmentation-Based Initialization for Steered Mixture of Experts Image Regression0
Boosting Code-Switching ASR with Mixture of Experts Enhanced Speech-Conditioned LLM0
Mean-field limit from general mixtures of experts to quantum neural networks0
Locking and Quacking: Stacking Bayesian model predictions by log-pooling and superposition0
Show:102550
← PrevPage 26 of 53Next →

No leaderboard results yet.