SOTAVerified

Mixture-of-Experts

Papers

Showing 501550 of 1312 papers

TitleStatusHype
A Novel Trustworthy Video Summarization Algorithm Through a Mixture of LoRA Experts0
Exploring Speaker Diarization with Mixture of Experts0
Advancing Robust Underwater Acoustic Target Recognition through Multi-task Learning and Multi-Gate Mixture-of-Experts0
IDEA: An Inverse Domain Expert Adaptation Based Active DNN IP Protection Method0
Identifying Shopping Intent in Product QA for Proactive Recommendations0
M6-T: Exploring Sparse Expert Models and Beyond0
Exploring Routing Strategies for Multilingual Mixture-of-Experts Models0
Imitation Learning from MPC for Quadrupedal Multi-Gait Control0
Imitation Learning from Observations: An Autoregressive Mixture of Experts Approach0
Improved Training of Mixture-of-Experts Language GANs0
A Hybrid Tensor-Expert-Data Parallelism Approach to Optimize Mixture-of-Experts Training0
CL-MoE: Enhancing Multimodal Large Language Model with Dual Momentum Mixture-of-Experts for Continual Visual Question Answering0
Exploring Domain Robust Lightweight Reward Models based on Router Mechanism0
Advancing MoE Efficiency: A Collaboration-Constrained Routing (C2R) Strategy for Better Expert Parallelism Design0
Little By Little: Continual Learning via Self-Activated Sparse Mixture-of-Rank Adaptive Learning0
CLIP-UP: A Simple and Efficient Mixture-of-Experts CLIP Training Recipe with Sparse Upcycling0
Exploiting Mixture-of-Experts Redundancy Unlocks Multimodal Generative Abilities0
A Novel Temporal Multi-Gate Mixture-of-Experts Approach for Vehicle Trajectory and Driving Intention Prediction0
Explainable data-driven modeling via mixture of experts: towards effective blending of grey and black-box models0
Ada-K Routing: Boosting the Efficiency of MoE-based LLMs0
Llama 3 Meets MoE: Efficient Upcycling0
LLaVA-CMoE: Towards Continual Mixture of Experts for Large Vision-Language Models0
Explainable Classifier for Malignant Lymphoma Subtyping via Cell Graph and Image Fusion0
Expert-Token Resonance: Redefining MoE Routing through Affinity-Driven Active Selection0
ClimateLLM: Efficient Weather Forecasting via Frequency-Aware Large Language Models0
Experts Weights Averaging: A New General Training Scheme for Vision Transformers0
CLER: Cross-task Learning with Expert Representation to Generalize Reading and Understanding0
A Novel Cluster Classify Regress Model Predictive Controller Formulation; CCR-MPC0
Advancing Expert Specialization for Better MoE0
ExpertRank: A Multi-level Coarse-grained Expert-based Listwise Ranking Loss0
ExpertRAG: Efficient RAG with Mixture of Experts -- Optimizing Context Retrieval for Adaptive LLM Responses0
CICADA: Cross-Domain Interpretable Coding for Anomaly Detection and Adaptation in Multivariate Time Series0
Expert Race: A Flexible Routing Strategy for Scaling Diffusion Transformer with Mixture of Experts0
ExpertFlow: Optimized Expert Activation and Token Allocation for Efficient Mixture-of-Experts Inference0
A Novel A.I Enhanced Reservoir Characterization with a Combined Mixture of Experts -- NVIDIA Modulus based Physics Informed Neural Operator Forward Model0
Expert Aggregation for Financial Forecasting0
Wonderful Matrices: More Efficient and Effective Architecture for Language Modeling Tasks0
EvoMoE: Expert Evolution in Mixture of Experts for Multimodal Large Language Models0
Advancing Enterprise Spatio-Temporal Forecasting Applications: Data Mining Meets Instruction Tuning of Language Models For Multi-modal Time Series Analysis in Low-Resource Settings0
AdaEnsemble: Learning Adaptively Sparse Structured Ensemble Network for Click-Through Rate Prediction0
EVLM: An Efficient Vision-Language Model for Visual Understanding0
EvidenceMoE: A Physics-Guided Mixture-of-Experts with Evidential Critics for Advancing Fluorescence Light Detection and Ranging in Scattering Media0
Every Sample Matters: Leveraging Mixture-of-Experts and High-Quality Data for Efficient and Accurate Code LLM0
Every FLOP Counts: Scaling a 300B Mixture-of-Experts LING LLM without Premium GPUs0
Non-asymptotic model selection in block-diagonal mixture of polynomial experts models0
3D Gaussian Splatting Data Compression with Mixture of Priors0
Every Expert Matters: Towards Effective Knowledge Distillation for Mixture-of-Experts Language Models0
EVE: Efficient Vision-Language Pre-training with Masked Prediction and Modality-Aware MoE0
Channel Gain Cartography via Mixture of Experts0
EVA: Mixture-of-Experts Semantic Variant Alignment for Compositional Zero-Shot Learning0
Show:102550
← PrevPage 11 of 27Next →

No leaderboard results yet.