SOTAVerified

Mixture-of-Experts

Papers

Showing 601650 of 1312 papers

TitleStatusHype
Efficient Residual Learning with Mixture-of-Experts for Universal Dexterous Grasping0
Breaking the gridlock in Mixture-of-Experts: Consistent and Efficient Algorithms0
Efficient Reflectance Capture with a Deep Gated Mixture-of-Experts0
Efficient Model Agnostic Approach for Implicit Neural Representation Based Arbitrary-Scale Image Super-Resolution0
Approximation Rates and VC-Dimension Bounds for (P)ReLU MLP Mixture of Experts0
An Effective Mixture-Of-Experts Approach For Code-Switching Speech Recognition Leveraging Encoder Disentanglement0
Many Hands Make Light Work: Task-Oriented Dialogue System with Module-Based Mixture-of-Experts0
Breaking Data Silos: Towards Open and Scalable Mobility Foundation Models via Generative Continual Learning0
Mix of Experts Language Model for Named Entity Recognition0
Mixture of Cluster-conditional LoRA Experts for Vision-language Instruction Tuning0
Mixture of Experts-augmented Deep Unfolding for Activity Detection in IRS-aided Systems0
Mixture of Experts in Large Language Models0
EfficientLLM: Efficiency in Large Language Models0
Efficient Large Scale Video Classification0
Branch-Train-MiX: Mixing Expert LLMs into a Mixture-of-Experts LLM0
Efficient Large Scale Language Modeling with Mixtures of Experts0
Efficient Language Modeling with Sparse all-MLP0
BrainNet-MoE: Brain-Inspired Mixture-of-Experts Learning for Neurological Disease Identification0
Anchoring to Exemplars for Training Mixture-of-Expert Cell Embeddings0
M^3TN: Multi-gate Mixture-of-Experts based Multi-valued Treatment Network for Uplift Modeling0
M2Restore: Mixture-of-Experts-based Mamba-CNN Fusion Framework for All-in-One Image Restoration0
M2R2: Mixture of Multi-Rate Residuals for Efficient Transformer Inference0
Efficient Deweather Mixture-of-Experts with Uncertainty-aware Feature-wise Linear Modulation0
M^2CD: A Unified MultiModal Framework for Optical-SAR Change Detection with Mixture of Experts and Self-Distillation0
Brain-Like Processing Pathways Form in Models With Heterogeneous Experts0
An Autonomous Network Orchestration Framework Integrating Large Language Models with Continual Reinforcement Learning0
AdaSpeech 3: Adaptive Text to Speech for Spontaneous Style0
MIRA: Medical Time Series Foundation Model for Real-World Health Data0
Lynx: Enabling Efficient MoE Inference through Dynamic Batch-Aware Expert Selection0
LSTM-based Mixture-of-Experts for Knowledge-Aware Dialogues0
Efficient Data Driven Mixture-of-Expert Extraction from Trained Networks0
M6-T: Exploring Sparse Expert Models and Beyond0
Machine learning based digital twin for dynamical systems with multiple time-scales0
LPT++: Efficient Training on Mixture of Long-tailed Experts0
Low-Rank Mixture-of-Experts for Continual Medical Image Segmentation0
MALoRA: Mixture of Asymmetric Low-Rank Adaptation for Enhanced Multi-Task Learning0
MANDARIN: Mixture-of-Experts Framework for Dynamic Delirium and Coma Prediction in ICU Patients: Development and Validation of an Acute Brain Dysfunction Prediction Model0
Boost Your NeRF: A Model-Agnostic Mixture of Experts Framework for High Quality and Efficient Rendering0
An Autonomous Negotiating Agent Framework with Reinforcement Learning Based Strategies and Adaptive Strategy Switching Mechanism0
Lory: Fully Differentiable Mixture-of-Experts for Autoregressive Language Model Pre-training0
LoRA-Switch: Boosting the Efficiency of Dynamic LLM Adapters via System-Algorithm Co-design0
Efficient Mixture-of-Expert for Video-based Driver State and Physiological Multi-task Estimation in Conditional Autonomous Driving0
Massively Multilingual Shallow Fusion with Large Language Models0
LoRA-Mixer: Coordinate Modular LoRA Experts Through Serial Attention Routing0
MAST-Pro: Dynamic Mixture-of-Experts for Adaptive Segmentation of Pan-Tumors with Knowledge-Driven Prompts0
Efficient and Effective Weight-Ensembling Mixture of Experts for Multi-Task Model Merging0
Adaptive Segmentation-Based Initialization for Steered Mixture of Experts Image Regression0
Boosting Code-Switching ASR with Mixture of Experts Enhanced Speech-Conditioned LLM0
Mean-field limit from general mixtures of experts to quantum neural networks0
Locking and Quacking: Stacking Bayesian model predictions by log-pooling and superposition0
Show:102550
← PrevPage 13 of 27Next →

No leaderboard results yet.