SOTAVerified

Mixture-of-Experts

Papers

Showing 651700 of 1312 papers

TitleStatusHype
Towards Adversarial Robustness of Model-Level Mixture-of-Experts Architectures for Semantic SegmentationCode0
Enhancing Healthcare Recommendation Systems with a Multimodal LLMs-based MOE Architecture0
Investigating Mixture of Experts in Dense Retrieval0
Llama 3 Meets MoE: Efficient Upcycling0
Mixture of Experts Meets Decoupled Message Passing: Towards General and Adaptive Node ClassificationCode0
Adaptive Prompting for Continual Relation Extraction: A Within-Task Variance Perspective0
MoE-CAP: Benchmarking Cost, Accuracy and Performance of Sparse Mixture-of-Experts Systems0
UniPaint: Unified Space-time Video Inpainting via Mixture-of-Experts0
An Entailment Tree Generation Approach for Multimodal Multi-Hop Question Answering with Mixture-of-Experts and Iterative Feedback Mechanism0
Towards 3D Acceleration for low-power Mixture-of-Experts and Multi-Head Attention Spiking Transformers0
Steps are all you need: Rethinking STEM Education with Prompt Engineering0
Convolutional Neural Networks and Mixture of Experts for Intrusion Detection in 5G Networks and beyond0
Yi-Lightning Technical Report0
HiMoE: Heterogeneity-Informed Mixture-of-Experts for Fair Spatial-Temporal Forecasting0
MQFL-FHE: Multimodal Quantum Federated Learning Framework with Fully Homomorphic Encryption0
Mixture of Experts for Node Classification0
LaVIDE: A Language-Vision Discriminator for Detecting Changes in Satellite Image with Map References0
On the effectiveness of discrete representations in sparse mixture of experts0
Mixture of Cache-Conditional Experts for Efficient Mobile Device Inference0
UOE: Unlearning One Expert Is Enough For Mixture-of-experts LLMS0
Mixture of Experts in Image Classification: What's the Sweet Spot?0
Complexity Experts are Task-Discriminative Learners for Any Image Restoration0
Enhancing Code-Switching ASR Leveraging Non-Peaky CTC Loss and Deep Language Posterior Injection0
H^3Fusion: Helpful, Harmless, Honest Fusion of Aligned LLMsCode0
LDACP: Long-Delayed Ad Conversions Prediction Model for Bidding Strategy0
MH-MoE: Multi-Head Mixture-of-Experts0
Lifelong Knowledge Editing for Vision Language Models with Low-Rank Mixture-of-Experts0
MERLOT: A Distilled LLM-based Mixture-of-Experts Framework for Scalable Encrypted Traffic Classification0
KAAE: Numerical Reasoning for Knowledge Graphs via Knowledge-aware Attributes Learning0
Ultra-Sparse Memory Network0
MoE-Lightning: High-Throughput MoE Inference on Memory-constrained GPUs0
Weakly-Supervised Multimodal Learning on MIMIC-CXRCode0
Sparse Upcycling: Inference Inefficient Finetuning0
Lynx: Enabling Efficient MoE Inference through Dynamic Batch-Aware Expert Selection0
Towards Vision Mixture of Experts for Wildlife Monitoring on the Edge0
PERFT: Parameter-Efficient Routed Fine-Tuning for Mixture-of-Expert Model0
Imitation Learning from Observations: An Autoregressive Mixture of Experts Approach0
Adaptive Conditional Expert Selection Network for Multi-domain Recommendation0
WDMoE: Wireless Distributed Mixture of Experts for Large Language Models0
NeKo: Toward Post Recognition Generative Correction Large Language Models with Task-Oriented Experts0
DA-MoE: Addressing Depth-Sensitivity in Graph-Level Analysis through Mixture of ExpertsCode0
Advancing Robust Underwater Acoustic Target Recognition through Multi-task Learning and Multi-Gate Mixture-of-Experts0
FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation0
HOBBIT: A Mixed Precision Expert Offloading System for Fast MoE Inference0
Facet-Aware Multi-Head Mixture-of-Experts Model for Sequential Recommendation0
RS-MoE: Mixture of Experts for Remote Sensing Image Captioning and Visual Question Answering0
PMoL: Parameter Efficient MoE for Preference Mixing of LLM Alignment0
MoNTA: Accelerating Mixture-of-Experts Training with Network-Traffc-Aware Parallel OptimizationCode0
MoE-I^2: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank DecompositionCode0
Stereo-Talker: Audio-driven 3D Human Synthesis with Prior-Guided Mixture-of-Experts0
Show:102550
← PrevPage 14 of 27Next →

No leaderboard results yet.