SOTAVerified

Mixture-of-Experts

Papers

Showing 11261150 of 1312 papers

TitleStatusHype
One Student Knows All Experts Know: From Sparse to Dense0
MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation0
Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners0
DeepSpeed-MoE: Advancing Mixture-of-Experts Inference and Training to Power Next-Generation AI ScaleCode0
Towards Lightweight Neural Animation : Exploration of Neural Network Pruning in Mixture of Experts-based Animation Models0
Combinations of Adaptive Filters0
Efficient Large Scale Language Modeling with Mixtures of Experts0
GLaM: Efficient Scaling of Language Models with Mixture-of-Experts0
Building a great multi-lingual teacher with sparsely-gated mixture of experts for speech recognition0
Specializing Versatile Skill Libraries using Local Mixture of ExpertsCode0
Anchoring to Exemplars for Training Mixture-of-Expert Cell Embeddings0
A Mixture of Expert Based Deep Neural Network for Improved ASR0
TAL: Two-stream Adaptive Learning for Generalizable Person Re-identification0
Expert Aggregation for Financial Forecasting0
SpeechMoE2: Mixture-of-Experts Model with Improved Routing0
Table-based Fact Verification with Self-adaptive Mixture of Experts0
MoEfication: Conditional Computation of Transformer Models for Efficient Inference0
StableMoE: Stable Routing Strategy for Mixture of Experts0
M6-T: Exploring Sparse Expert Models and Beyond0
SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization0
Elucidating Robust Learning with Uncertainty-Aware Corruption Pattern EstimationCode0
RTM Super Learner Results at Quality Estimation Task0
Polynomial-Spline Neural Networks with Exact Integrals0
P-Adapters: Robustly Extracting Factual Information from Language Models with Diverse Prompts0
Simple or Complex? Complexity-Controllable Question Generation with Soft Templates and Deep Mixture of Experts Model0
Show:102550
← PrevPage 46 of 53Next →

No leaderboard results yet.