SOTAVerified

Mixture-of-Experts

Papers

Showing 601625 of 1312 papers

TitleStatusHype
MoExtend: Tuning New Experts for Modality and Task ExtensionCode1
Mixture-of-Noises Enhanced Forgery-Aware Predictor for Multi-Face Manipulation Detection and Localization0
HMDN: Hierarchical Multi-Distribution Network for Click-Through Rate Prediction0
Multimodal Fusion and Coherence Modeling for Video Topic Segmentation0
PMoE: Progressive Mixture of Experts with Asymmetric Transformer for Continual Learning0
MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts0
Distribution Learning for Molecular Regression0
Time series forecasting with high stakes: A field study of the air cargo industry0
Mixture of Nested Experts: Adaptive Processing of Visual TokensCode0
Mixture of Modular Experts: Distilling Knowledge from a Multilingual Teacher into Specialized Modular Language ModelsCode0
MOoSE: Multi-Orientation Sharing Experts for Open-set Scene Text RecognitionCode0
Wolf: Captioning Everything with a World Summarization Framework0
Dynamic Language Group-Based MoE: Enhancing Code-Switching Speech Recognition with Hierarchical RoutingCode1
How Lightweight Can A Vision Transformer Be0
Exploring Domain Robust Lightweight Reward Models based on Router Mechanism0
Wonderful Matrices: More Efficient and Effective Architecture for Language Modeling Tasks0
M4: Multi-Proxy Multi-Gate Mixture of Experts Network for Multiple Instance Learning in Histopathology Image AnalysisCode1
Stretching Each Dollar: Diffusion Training from Scratch on a Micro-BudgetCode5
Norface: Improving Facial Expression Analysis by Identity NormalizationCode1
EEGMamba: Bidirectional State Space Model with Mixture of Experts for EEG Multi-task Classification0
Mixture of Experts with Mixture of Precisions for Tuning Quality of Service0
EVLM: An Efficient Vision-Language Model for Visual Understanding0
Mixture of Experts based Multi-task Supervise Learning from Crowds0
Discussion: Effective and Interpretable Outcome Prediction by Training Sparse Mixtures of Linear Experts0
MoE-DiffIR: Task-customized Diffusion Priors for Universal Compressed Image Restoration0
Show:102550
← PrevPage 25 of 53Next →

No leaderboard results yet.