SOTAVerified

Mixture-of-Experts

Papers

Showing 11511200 of 1312 papers

TitleStatusHype
KDExplainer: A Task-oriented Attention Model for Explaining Knowledge Distillation0
Knowledge-in-Context: Towards Knowledgeable Semi-Parametric Language Models0
LaDiMo: Layer-wise Distillation Inspired MoEfier0
Lancet: Accelerating Mixture-of-Experts Training via Whole Graph Computation-Communication Overlapping0
Language-driven All-in-one Adverse Weather Removal0
Large-Scale YouTube-8M Video Understanding with Deep Neural Networks0
La-SoftMoE CLIP for Unified Physical-Digital Face Attack Detection0
LaVIDE: A Language-Vision Discriminator for Detecting Changes in Satellite Image with Map References0
Lazarus: Resilient and Elastic Training of Mixture-of-Experts Models with Adaptive Expert Placement0
LDACP: Long-Delayed Ad Conversions Prediction Model for Bidding Strategy0
Learning Factored Representations in a Deep Mixture of Experts0
Learning Heterogeneous Tissues with Mixture of Experts for Gigapixel Whole Slide Images0
Learning in Gated Neural Networks0
Learning Large-scale Universal User Representation with Sparse Mixture of Experts0
Learning More Generalized Experts by Merging Experts in Mixture-of-Experts0
Learning Sparse Mixture of Experts for Visual Question Answering0
Learning to Compose Topic-Aware Mixture of Experts for Zero-Shot Video Captioning0
Learning to Ground VLMs without Forgetting0
LEMoE: Advanced Mixture of Experts Adaptor for Lifelong Model Editing of Large Language Models0
Leveraging Mixture of Experts for Improved Speech Deepfake Detection0
Leveraging MoE-based Large Language Model for Zero-Shot Multi-Task Semantic Communication0
Leveraging Pre-Trained Models for Multimodal Class-Incremental Learning under Adaptive Fusion0
Lifelong Evolution: Collaborative Learning between Large and Small Language Models for Continuous Emergent Fake News Detection0
Lifelong Knowledge Editing for Vision Language Models with Low-Rank Mixture-of-Experts0
Lifelong Language Pretraining with Distribution-Specialized Experts0
Little By Little: Continual Learning via Self-Activated Sparse Mixture-of-Rank Adaptive Learning0
Llama 3 Meets MoE: Efficient Upcycling0
LLaVA-CMoE: Towards Continual Mixture of Experts for Large Vision-Language Models0
LLaVA-MoLE: Sparse Mixture of LoRA Experts for Mitigating Data Conflicts in Instruction Finetuning MLLMs0
LLM4WM: Adapting LLM for Wireless Multi-Tasking0
LLM-Based Routing in Mixture of Experts: A Novel Framework for Trading0
Load Balancing Mixture of Experts with Similarity Preserving Routers0
Locking and Quacking: Stacking Bayesian model predictions by log-pooling and superposition0
LoRA-Mixer: Coordinate Modular LoRA Experts Through Serial Attention Routing0
LoRA-Switch: Boosting the Efficiency of Dynamic LLM Adapters via System-Algorithm Co-design0
Lory: Fully Differentiable Mixture-of-Experts for Autoregressive Language Model Pre-training0
Low-Rank Mixture-of-Experts for Continual Medical Image Segmentation0
LPT++: Efficient Training on Mixture of Long-tailed Experts0
LSTM-based Mixture-of-Experts for Knowledge-Aware Dialogues0
Lynx: Enabling Efficient MoE Inference through Dynamic Batch-Aware Expert Selection0
M^2CD: A Unified MultiModal Framework for Optical-SAR Change Detection with Mixture of Experts and Self-Distillation0
M2R2: Mixture of Multi-Rate Residuals for Efficient Transformer Inference0
M2Restore: Mixture-of-Experts-based Mamba-CNN Fusion Framework for All-in-One Image Restoration0
M^3TN: Multi-gate Mixture-of-Experts based Multi-valued Treatment Network for Uplift Modeling0
M6-T: Exploring Sparse Expert Models and Beyond0
Machine learning based digital twin for dynamical systems with multiple time-scales0
MALoRA: Mixture of Asymmetric Low-Rank Adaptation for Enhanced Multi-Task Learning0
MANDARIN: Mixture-of-Experts Framework for Dynamic Delirium and Coma Prediction in ICU Patients: Development and Validation of an Acute Brain Dysfunction Prediction Model0
Many Hands Make Light Work: Task-Oriented Dialogue System with Module-Based Mixture-of-Experts0
Massively Multilingual Shallow Fusion with Large Language Models0
Show:102550
← PrevPage 24 of 27Next →

No leaderboard results yet.