SOTAVerified

Mixture-of-Experts

Papers

Showing 126150 of 1312 papers

TitleStatusHype
MiniDrive: More Efficient Vision-Language Models with Multi-Level 2D Features as Text Tokens for Autonomous DrivingCode2
CNMBERT: A Model for Converting Hanyu Pinyin Abbreviations to Chinese CharactersCode2
CLIP-MoE: Towards Building Mixture of Experts for CLIP with Diversified Multiplet UpcyclingCode2
MC-MoE: Mixture Compressor for Mixture-of-Experts LLMs Gains MoreCode2
Object Detection using Event Camera: A MoE Heat Conduction based Detector and A New Benchmark DatasetCode2
Superposition in Transformers: A Novel Way of Building Mixture of ExpertsCode2
M4: Multi-Proxy Multi-Gate Mixture of Experts Network for Multiple Instance Learning in Histopathology Image AnalysisCode1
M^3ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task Learning with Model-Accelerator Co-designCode1
M^4oE: A Foundation Model for Medical Multimodal Image Segmentation with Mixture of ExpertsCode1
M3oE: Multi-Domain Multi-Task Mixture-of Experts Recommendation FrameworkCode1
LMHaze: Intensity-aware Image Dehazing with a Large-scale Multi-intensity Real Haze DatasetCode1
M3-Jepa: Multimodal Alignment via Multi-directional MoE based on the JEPA frameworkCode1
LOLA -- An Open-Source Massively Multilingual Large Language ModelCode1
LLMBind: A Unified Modality-Task Integration FrameworkCode1
PAD-Net: An Efficient Framework for Dynamic NetworksCode1
LLMCarbon: Modeling the end-to-end Carbon Footprint of Large Language ModelsCode1
Long-Tailed Visual Recognition via Self-Heterogeneous Integration with Knowledge ExcavationCode1
Making Neural Networks Interpretable with Attribution: Application to Implicit Signals PredictionCode1
LITE: Modeling Environmental Ecosystems with Multimodal Large Language ModelsCode1
AlphaLoRA: Assigning LoRA Experts Based on Layer Training QualityCode1
A Time Series is Worth Five Experts: Heterogeneous Mixture of Experts for Traffic Flow PredictionCode1
Learning to Skip the Middle Layers of TransformersCode1
LIBMoE: A Library for comprehensive benchmarking Mixture of Experts in Large Language ModelsCode1
ChatVLA: Unified Multimodal Understanding and Robot Control with Vision-Language-Action ModelCode1
Parameter Efficient Adaptation for Image Restoration with Heterogeneous Mixture-of-ExpertsCode1
Show:102550
← PrevPage 6 of 53Next →

No leaderboard results yet.