SOTAVerified

Mixture-of-Experts

Papers

Showing 151200 of 1312 papers

TitleStatusHype
PICO: Secure Transformers via Robust Prompt Isolation and Cybersecurity Oversight0
NoEsis: Differentially Private Knowledge Transfer in Modular LLM Adaptation0
Unveiling the Hidden: Movie Genre and User Bias in Spoiler DetectionCode0
BadMoE: Backdooring Mixture-of-Experts LLMs via Optimizing Routing Triggers and Infecting Dormant Experts0
Manifold Induced Biases for Zero-shot and Few-shot Detection of Generated ImagesCode1
MoE Parallel Folding: Heterogeneous Parallelism Mappings for Efficient Large-Scale MoE Model Training with Megatron Core0
Distribution-aware Forgetting Compensation for Exemplar-Free Lifelong Person Re-identificationCode1
Multi-Type Context-Aware Conversational Recommender Systems via Mixture-of-Experts0
HAECcity: Open-Vocabulary Scene Understanding of City-Scale Point Clouds with Superpoint Graph Clustering0
D^2MoE: Dual Routing and Dynamic Scheduling for Efficient On-Device MoE-based LLM Serving0
Trend Filtered Mixture of Experts for Automated Gating of High-Frequency Flow Cytometry Data0
Dense Backpropagation Improves Training for Sparse Mixture-of-ExpertsCode1
Unveiling Hidden Collaboration within Mixture-of-Experts in Large Language Models0
Plasticity-Aware Mixture of Experts for Learning Under QoE Shifts in Adaptive Video Streaming0
Mixture-of-Shape-Experts (MoSE): End-to-End Shape Dictionary Framework to Prompt SAM for Generalizable Medical Segmentation0
MoE-Lens: Towards the Hardware Limit of High-Throughput MoE LLM Serving Under Resource Constraints0
RouterKT: Mixture-of-Experts for Knowledge TracingCode0
Regularized infill criteria for multi-objective Bayesian optimization with application to aircraft design0
Seed1.5-Thinking: Advancing Superb Reasoning Models with Reinforcement Learning0
Kimi-VL Technical ReportCode5
Scaling Laws for Native Multimodal Models Scaling Laws for Native Multimodal Models0
Cluster-Driven Expert Pruning for Mixture-of-Experts Large Language ModelsCode0
C3PO: Critical-Layer, Core-Expert, Collaborative Pathway Optimization for Test-Time Expert Re-MixingCode1
Adaptive Detection of Fast Moving Celestial Objects Using a Mixture of Experts and Physical-Inspired Neural Network0
Holistic Capability Preservation: Towards Compact Yet Comprehensive Reasoning Models0
FedMerge: Federated Personalization via Model Merging0
MoEDiff-SR: Mixture of Experts-Guided Diffusion Model for Region-Adaptive MRI Super-ResolutionCode1
Finding Fantastic Experts in MoEs: A Unified Study for Expert Dropping Strategies and Observations0
HybriMoE: Hybrid CPU-GPU Scheduling and Cache Management for Efficient MoE InferenceCode2
RingMoE: Mixture-of-Modality-Experts Multi-Modal Foundation Models for Universal Remote Sensing Image Interpretation0
HeterMoE: Efficient Training of Mixture-of-Experts Models on Heterogeneous GPUs0
MiLo: Efficient Quantized MoE Inference with Mixture of Low-Rank CompensatorsCode1
MegaScale-Infer: Serving Mixture-of-Experts at Scale with Disaggregated Expert Parallelism0
Advancing MoE Efficiency: A Collaboration-Constrained Routing (C2R) Strategy for Better Expert Parallelism Design0
A Unified Virtual Mixture-of-Experts Framework:Enhanced Inference and Hallucination Mitigation in Single-Model System0
Detecting Financial Fraud with Hybrid Deep Learning: A Mix-of-Experts Approach to Sequential and Anomalous Patterns0
DynMoLE: Boosting Mixture of LoRA Experts Fine-Tuning with a Hybrid Routing MechanismCode0
Unimodal-driven Distillation in Multimodal Emotion Recognition with Dynamic Fusion0
Mixture of Routers0
Sparse Mixture of Experts as Unified Competitive Learning0
S2MoE: Robust Sparse Mixture of Experts via Stochastic Learning0
Beyond Standard MoE: Mixture of Latent Experts for Resource-Efficient Language Models0
Exploiting Mixture-of-Experts Redundancy Unlocks Multimodal Generative Abilities0
RocketPPA: Code-Level Power, Performance, and Area Prediction via LLM and Mixture of Experts0
LLaVA-CMoE: Towards Continual Mixture of Experts for Large Vision-Language Models0
iMedImage Technical Report0
A multi-scale lithium-ion battery capacity prediction using mixture of experts and patch-based MLPCode0
Reasoning Beyond Limits: Advances and Open Problems for LLMs0
Enhancing Multi-modal Models with Heterogeneous MoE Adapters for Fine-tuning0
Optimal Scaling Laws for Efficiency Gains in a Theoretical Transformer-Augmented Sectional MoE Framework0
Show:102550
← PrevPage 4 of 27Next →

No leaderboard results yet.