| MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation | Apr 15, 2022 | Knowledge DistillationMixture-of-Experts | CodeCode Available | 1 |
| Mixture-of-experts VAEs can disregard variation in surjective multimodal data | Apr 11, 2022 | Mixture-of-Experts | —Unverified | 0 |
| 3M: Multi-loss, Multi-path and Multi-level Neural Networks for speech recognition | Apr 7, 2022 | Mixture-of-Expertsspeech-recognition | CodeCode Available | 1 |
| Learning to Adapt Clinical Sequences with Residual Mixture of Experts | Apr 6, 2022 | Mixture-of-Experts | CodeCode Available | 0 |
| Combining Spectral and Self-Supervised Features for Low Resource Speech Recognition and Translation | Apr 5, 2022 | Automatic Speech RecognitionAutomatic Speech Recognition (ASR) | —Unverified | 0 |
| On the Adaptation to Concept Drift for CTR Prediction | Apr 1, 2022 | Click-Through Rate PredictionIncremental Learning | —Unverified | 0 |
| Efficient Reflectance Capture with a Deep Gated Mixture-of-Experts | Mar 29, 2022 | DecoderMixture-of-Experts | —Unverified | 0 |
| Efficient and Degradation-Adaptive Network for Real-World Image Super-Resolution | Mar 27, 2022 | Image Super-ResolutionMixture-of-Experts | CodeCode Available | 1 |
| Build a Robust QA System with Transformer-based Mixture of Experts | Mar 20, 2022 | Data AugmentationMixture-of-Experts | CodeCode Available | 0 |
| Efficient Language Modeling with Sparse all-MLP | Mar 14, 2022 | AllCommon Sense Reasoning | —Unverified | 0 |
| SummaReranker: A Multi-Task Mixture-of-Experts Re-ranking Framework for Abstractive Summarization | Mar 13, 2022 | Abstractive Text SummarizationDocument Summarization | CodeCode Available | 1 |
| SkillNet-NLU: A Sparsely Activated Model for General-Purpose Natural Language Understanding | Mar 7, 2022 | Language ModellingMasked Language Modeling | —Unverified | 0 |
| Parameter-Efficient Mixture-of-Experts Architecture for Pre-trained Language Models | Mar 2, 2022 | Language ModelingLanguage Modelling | CodeCode Available | 1 |
| Functional mixture-of-experts for classification | Feb 28, 2022 | ClassificationMixture-of-Experts | —Unverified | 0 |
| Mixture-of-Experts with Expert Choice Routing | Feb 18, 2022 | Mixture-of-Experts | —Unverified | 0 |
| ST-MoE: Designing Stable and Transferable Sparse Expert Models | Feb 17, 2022 | ARCCommon Sense Reasoning | CodeCode Available | 3 |
| A Survey on Dynamic Neural Networks for Natural Language Processing | Feb 15, 2022 | Dynamic neural networksMixture-of-Experts | —Unverified | 0 |
| Physics-Guided Problem Decomposition for Scaling Deep Learning of High-dimensional Eigen-Solvers: The Case of Schrödinger's Equation | Feb 12, 2022 | Mixture-of-ExpertsProblem Decomposition | —Unverified | 0 |
| One Student Knows All Experts Know: From Sparse to Dense | Jan 26, 2022 | AllKnowledge Distillation | —Unverified | 0 |
| Sparsely Activated Mixture-of-Experts are Robust Multi-Task Learners | Jan 16, 2022 | Mixture-of-ExpertsMulti-Task Learning | —Unverified | 0 |
| MoEBERT: from BERT to Mixture-of-Experts via Importance-Guided Adaptation | Jan 16, 2022 | Knowledge DistillationMixture-of-Experts | —Unverified | 0 |
| DeepSpeed-MoE: Advancing Mixture-of-Experts Inference and Training to Power Next-Generation AI Scale | Jan 14, 2022 | DecoderMixture-of-Experts | CodeCode Available | 0 |
| Towards Lightweight Neural Animation : Exploration of Neural Network Pruning in Mixture of Experts-based Animation Models | Jan 11, 2022 | Mixture-of-ExpertsNetwork Pruning | —Unverified | 0 |
| MDFEND: Multi-domain Fake News Detection | Jan 4, 2022 | Fake News DetectionMixture-of-Experts | CodeCode Available | 2 |
| EvoMoE: An Evolutional Mixture-of-Experts Training Framework via Dense-To-Sparse Gate | Dec 29, 2021 | Language ModelingLanguage Modelling | CodeCode Available | 1 |