SOTAVerified

Mixture-of-Experts

Papers

Showing 11011125 of 1312 papers

TitleStatusHype
Combinations of Adaptive Filters0
Efficient Large Scale Language Modeling with Mixtures of Experts0
Mimic Embedding via Adaptive Aggregation: Learning Generalizable Person Re-identificationCode1
GLaM: Efficient Scaling of Language Models with Mixture-of-Experts0
Building a great multi-lingual teacher with sparsely-gated mixture of experts for speech recognition0
Specializing Versatile Skill Libraries using Local Mixture of ExpertsCode0
Anchoring to Exemplars for Training Mixture-of-Expert Cell Embeddings0
A Mixture of Expert Based Deep Neural Network for Improved ASR0
TAL: Two-stream Adaptive Learning for Generalizable Person Re-identification0
Expert Aggregation for Financial Forecasting0
SpeechMoE2: Mixture-of-Experts Model with Improved Routing0
M6-T: Exploring Sparse Expert Models and Beyond0
StableMoE: Stable Routing Strategy for Mixture of Experts0
Table-based Fact Verification with Self-adaptive Mixture of Experts0
SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization0
MoEfication: Conditional Computation of Transformer Models for Efficient Inference0
Elucidating Robust Learning with Uncertainty-Aware Corruption Pattern EstimationCode0
RTM Super Learner Results at Quality Estimation Task0
Unsupervised Foreground Extraction via Deep Region CompetitionCode1
Polynomial-Spline Neural Networks with Exact Integrals0
P-Adapters: Robustly Extracting Factual Information from Language Models with Diverse Prompts0
Simple or Complex? Complexity-Controllable Question Generation with Soft Templates and Deep Mixture of Experts Model0
HydraSum: Disentangling Stylistic Features in Text Summarization using Multi-Decoder ModelsCode1
Taming Sparsely Activated Transformer with Stochastic ExpertsCode1
Sparse MoEs meet Efficient EnsemblesCode1
Show:102550
← PrevPage 45 of 53Next →

No leaderboard results yet.