SOTAVerified

Mixture-of-Experts

Papers

Showing 451500 of 1312 papers

TitleStatusHype
Complexity Experts are Task-Discriminative Learners for Any Image Restoration0
A Survey on Model MoErging: Recycling and Routing Among Specialized Experts for Collaborative Learning0
Finding Fantastic Experts in MoEs: A Unified Study for Expert Dropping Strategies and Observations0
Graph Mixture of Experts and Memory-augmented Routers for Multivariate Time Series Anomaly Detection0
Filtered not Mixed: Stochastic Filtering-Based Online Gating for Mixture of Large Language Models0
A Review of DeepSeek Models' Key Innovative Techniques0
AdaMV-MoE: Adaptive Multi-Task Vision Mixture-of-Experts0
GRIN: GRadient-INformed MoE0
Language-driven All-in-one Adverse Weather Removal0
A Theoretical View on Sparsely Activated Networks0
Large-Scale YouTube-8M Video Understanding with Deep Neural Networks0
Learning More Generalized Experts by Merging Experts in Mixture-of-Experts0
Affect in Tweets Using Experts Model0
Learning to Specialize: Joint Gating-Expert Training for Adaptive MoEs in Decentralized Settings0
FedMoE: Personalized Federated Learning via Heterogeneous Mixture of Experts0
FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation0
KAT-V1: Kwai-AutoThink Technical Report0
HDformer: A Higher Dimensional Transformer for Diabetes Detection Utilizing Long Range Vascular Signals0
HeterMoE: Efficient Training of Mixture-of-Experts Models on Heterogeneous GPUs0
FedMerge: Federated Personalization via Model Merging0
A Provably Effective Method for Pruning Experts in Fine-tuned Sparse Mixture-of-Experts0
Heuristic-Informed Mixture of Experts for Link Prediction in Multilayer Networks0
KDExplainer: A Task-oriented Attention Model for Explaining Knowledge Distillation0
Decoding Knowledge Attribution in Mixture-of-Experts: A Framework of Basic-Refinement Collaboration and Efficiency Analysis0
Knowledge-in-Context: Towards Knowledgeable Semi-Parametric Language Models0
Federated Mixture of Experts0
Hierarchical Mixture-of-Experts Model for Large-Scale Gaussian Process Regression0
Deep Gaussian Covariance Network0
Federated learning using mixture of experts0
Combining Spectral and Self-Supervised Features for Low Resource Speech Recognition and Translation0
FEAMOE: Fair, Explainable and Adaptive Mixture of Experts0
Combining Parametric and Nonparametric Models for Off-Policy Evaluation0
FaVChat: Unlocking Fine-Grained Facail Video Understanding with Multimodal Large Language Models0
HMOE: Hypernetwork-based Mixture of Experts for Domain Generalization0
Combinations of Adaptive Filters0
Holistic Capability Preservation: Towards Compact Yet Comprehensive Reasoning Models0
Aphasic Speech Recognition using a Mixture of Speech Intelligibility Experts0
A Dynamic Approach to Stock Price Prediction: Comparing RNN and Mixture of Experts Models Across Different Volatility Profiles0
LaDiMo: Layer-wise Distillation Inspired MoEfier0
How Do Consumers Really Choose: Exposing Hidden Preferences with the Mixture of Experts Model0
La-SoftMoE CLIP for Unified Physical-Digital Face Attack Detection0
How Lightweight Can A Vision Transformer Be0
Learning Heterogeneous Tissues with Mixture of Experts for Gigapixel Whole Slide Images0
Lifelong Knowledge Editing for Vision Language Models with Low-Rank Mixture-of-Experts0
Hunyuan-TurboS: Advancing Large Language Models through Mamba-Transformer Synergy and Adaptive Chain-of-Thought0
Faster MoE LLM Inference for Extremely Large Models0
Faster Language Models with Better Multi-Token Prediction Using Tensor Decomposition0
CoCoAFusE: Beyond Mixtures of Experts via Model Fusion0
Fast, Differentiable and Sparse Top-k: a Convex Analysis Perspective0
An Unsupervised Domain Adaptation Method for Locating Manipulated Region in partially fake Audio0
Show:102550
← PrevPage 10 of 27Next →

No leaderboard results yet.