SOTAVerified

Mixture-of-Experts

Papers

Showing 701750 of 1312 papers

TitleStatusHype
Divide, Conquer, and Combine: Mixture of Semantic-Independent Experts for Zero-Shot Dialogue State Tracking0
Double Deep Q-Learning in Opponent Modeling0
Double-Stage Feature-Level Clustering-Based Mixture of Experts Framework0
Double-Wing Mixture of Experts for Streaming Recommendations0
DriveMoE: Mixture-of-Experts for Vision-Language-Action Model in End-to-End Autonomous Driving0
Dropout Regularization in Hierarchical Mixture of Experts0
Drop-Upcycling: Training Sparse Mixture of Experts with Partial Re-initialization0
DSMoE: Matrix-Partitioned Experts with Dynamic Routing for Computation-Efficient Dense LLMs0
DualComp: End-to-End Learning of a Unified Dual-Modality Lossless Compressor0
Duplex: A Device for Large Language Models with Mixture of Experts, Grouped Query Attention, and Continuous Batching0
Each Rank Could be an Expert: Single-Ranked Mixture of Experts LoRA for Multi-Task Learning0
EC-DIT: Scaling Diffusion Transformers with Adaptive Expert-Choice Routing0
ECG-EmotionNet: Nested Mixture of Expert (NMoE) Adaptation of ECG-Foundation Model for Driver Emotion Recognition0
Edge-Aware Autoencoder Design for Real-Time Mixture-of-Experts Image Compression0
EEGMamba: Bidirectional State Space Model with Mixture of Experts for EEG Multi-task Classification0
Efficient and Effective Weight-Ensembling Mixture of Experts for Multi-Task Model Merging0
Efficient Data Driven Mixture-of-Expert Extraction from Trained Networks0
Efficient Deweather Mixture-of-Experts with Uncertainty-aware Feature-wise Linear Modulation0
Efficient Language Modeling with Sparse all-MLP0
Efficient Large Scale Language Modeling with Mixtures of Experts0
Efficient Large Scale Video Classification0
EfficientLLM: Efficiency in Large Language Models0
Efficient Mixture-of-Expert for Video-based Driver State and Physiological Multi-task Estimation in Conditional Autonomous Driving0
Efficient Model Agnostic Approach for Implicit Neural Representation Based Arbitrary-Scale Image Super-Resolution0
Efficient Reflectance Capture with a Deep Gated Mixture-of-Experts0
Efficient Residual Learning with Mixture-of-Experts for Universal Dexterous Grasping0
Efficient Training of Large-Scale AI Models Through Federated Mixture-of-Experts: A System-Level Approach0
eMoE: Task-aware Memory Efficient Mixture-of-Experts-Based (MoE) Model Inference0
ENACT-Heart -- ENsemble-based Assessment Using CNN and Transformer on Heart Sounds0
Enhancing Code-Switching ASR Leveraging Non-Peaky CTC Loss and Deep Language Posterior Injection0
Enhancing Code-Switching Speech Recognition with LID-Based Collaborative Mixture of Experts Model0
Enhancing Generalization in Sparse Mixture of Experts Models: The Case for Increased Expert Activation in Compositional Tasks0
Enhancing Healthcare Recommendation Systems with a Multimodal LLMs-based MOE Architecture0
Enhancing Multimodal Continual Instruction Tuning with BranchLoRA0
Enhancing Multi-modal Models with Heterogeneous MoE Adapters for Fine-tuning0
Enhancing the "Immunity" of Mixture-of-Experts Networks for Adversarial Defense0
Ensemble Learning for Large Language Models in Text and Code Generation: A Survey0
EPS-MoE: Expert Pipeline Scheduler for Cost-Efficient MoE Inference0
Evaluating Expert Contributions in a MoE LLM for Quiz-Based Tasks0
EVA: Mixture-of-Experts Semantic Variant Alignment for Compositional Zero-Shot Learning0
EVE: Efficient Vision-Language Pre-training with Masked Prediction and Modality-Aware MoE0
Every Expert Matters: Towards Effective Knowledge Distillation for Mixture-of-Experts Language Models0
Every FLOP Counts: Scaling a 300B Mixture-of-Experts LING LLM without Premium GPUs0
Every Sample Matters: Leveraging Mixture-of-Experts and High-Quality Data for Efficient and Accurate Code LLM0
EvidenceMoE: A Physics-Guided Mixture-of-Experts with Evidential Critics for Advancing Fluorescence Light Detection and Ranging in Scattering Media0
EVLM: An Efficient Vision-Language Model for Visual Understanding0
EvoMoE: Expert Evolution in Mixture of Experts for Multimodal Large Language Models0
Expert Aggregation for Financial Forecasting0
ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference0
Expert Race: A Flexible Routing Strategy for Scaling Diffusion Transformer with Mixture of Experts0
Show:102550
← PrevPage 15 of 27Next →

No leaderboard results yet.