SOTAVerified

Mixture-of-Experts

Papers

Showing 351400 of 1312 papers

TitleStatusHype
Not Eliminate but Aggregate: Post-Hoc Control over Mixture-of-Experts to Address Shortcut Shifts in Natural Language UnderstandingCode0
Non-Normal Mixtures of ExpertsCode0
Octavius: Mitigating Task Interference in MLLMs via LoRA-MoECode0
Named Entity and Relation Extraction with Multi-Modal RetrievalCode0
AskChart: Universal Chart Understanding through Textual EnhancementCode0
Nesti-Net: Normal Estimation for Unstructured 3D Point Clouds using Convolutional Neural NetworksCode0
Multi-view Contrastive Learning for Entity Typing over Knowledge GraphsCode0
Multi-Source Domain Adaptation with Mixture of ExpertsCode0
ASEM: Enhancing Empathy in Chatbot through Attention-based Sentiment and Emotion ModelingCode0
Multi-modal Collaborative Optimization and Expansion Network for Event-assisted Single-eye Expression RecognitionCode0
Condensing Multilingual Knowledge with Lightweight Language-Specific ModulesCode0
A Gaussian Process-based Streaming Algorithm for Prediction of Time Series With Regimes and OutliersCode0
Multimodal Cultural Safety: Evaluation Frameworks and Alignment StrategiesCode0
AdaMV-MoE: Adaptive Multi-Task Vision Mixture-of-ExpertsCode0
A Gated Residual Kolmogorov-Arnold Networks for Mixtures of ExpertsCode0
Completed Feature Disentanglement Learning for Multimodal MRIs AnalysisCode0
MoVEInt: Mixture of Variational Experts for Learning Human-Robot Interactions from DemonstrationsCode0
Multimodal Fusion Strategies for Mapping Biophysical Landscape FeaturesCode0
CompeteSMoE -- Statistically Guaranteed Mixture of Experts Training via CompetitionCode0
MOoSE: Multi-Orientation Sharing Experts for Open-set Scene Text RecognitionCode0
MoRE-Brain: Routed Mixture of Experts for Interpretable and Generalizable Cross-Subject fMRI Visual DecodingCode0
pFedMoE: Data-Level Personalization with Mixture of Experts for Model-Heterogeneous Personalized Federated LearningCode0
FEDKIM: Adaptive Federated Knowledge Injection into Medical Foundation ModelsCode0
MoNTA: Accelerating Mixture-of-Experts Training with Network-Traffc-Aware Parallel OptimizationCode0
More Experts Than Galaxies: Conditionally-overlapping Experts With Biologically-Inspired Fixed RoutingCode0
Combining Spectral and Self-Supervised Features for Low Resource Speech Recognition and TranslationCode0
MoLEx: Mixture of Layer Experts for Finetuning with Sparse UpcyclingCode0
CoLA: Collaborative Low-Rank AdaptationCode0
Fast filtering of non-Gaussian models using Amortized Optimal Transport MapsCode0
Mol-MoE: Training Preference-Guided Routers for Molecule GenerationCode0
Mosaic: Data-Free Knowledge Distillation via Mixture-of-Experts for Heterogeneous Distributed EnvironmentsCode0
On-Device Collaborative Language Modeling via a Mixture of Generalists and SpecialistsCode0
MoELoRA: Contrastive Learning Guided Mixture of Experts on Parameter-Efficient Fine-Tuning for Large Language ModelsCode0
FactorLLM: Factorizing Knowledge via Mixture of Experts for Large Language ModelsCode0
MoE-I^2: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank DecompositionCode0
MoE-LPR: Multilingual Extension of Large Language Models through Mixture-of-Experts with Language Priors RoutingCode0
Extreme Classification in Log Memory using Count-Min Sketch: A Case Study of Amazon Search with 50M ProductsCode0
Cluster-Driven Expert Pruning for Mixture-of-Experts Large Language ModelsCode0
Exploring Model Consensus to Generate Translation ParaphrasesCode0
A Hybrid Tensor-Expert-Data Parallelism Approach to Optimize Mixture-of-Experts TrainingCode0
Modality-Independent Brain Lesion Segmentation with Privacy-aware Continual LearningCode0
Modeling Task Relationships in Multi-task Learning with Multi-gate Mixture-of-ExpertsCode0
Exploiting Activation Sparsity with Dense to Dynamic-k Mixture-of-Experts ConversionCode0
MLP-KAN: Unifying Deep Representation and Function LearningCode0
MoE-MLoRA for Multi-Domain CTR Prediction: Efficient Adaptation with Expert SpecializationCode0
CompeteSMoE - Effective Training of Sparse Mixture of Experts via CompetitionCode0
Mixture of Modular Experts: Distilling Knowledge from a Multilingual Teacher into Specialized Modular Language ModelsCode0
Mixture-of-LoRAs: An Efficient Multitask Tuning for Large Language ModelsCode0
Mixture of Nested Experts: Adaptive Processing of Visual TokensCode0
Mixture of Link Predictors on GraphsCode0
Show:102550
← PrevPage 8 of 27Next →

No leaderboard results yet.