| MoEC: Mixture of Expert Clusters | Jul 19, 2022 | Machine TranslationMixture-of-Experts | —Unverified | 0 |
| Learning Large-scale Universal User Representation with Sparse Mixture of Experts | Jul 11, 2022 | Mixture-of-Experts | —Unverified | 0 |
| No Language Left Behind: Scaling Human-Centered Machine Translation | Jul 11, 2022 | Machine TranslationMixture-of-Experts | CodeCode Available | 2 |
| DeepSpeed Inference: Enabling Efficient Inference of Transformer Models at Unprecedented Scale | Jun 30, 2022 | CPUGPU | CodeCode Available | 4 |
| RoME: Role-aware Mixture-of-Expert Transformer for Text-to-Video Retrieval | Jun 26, 2022 | Mixture-of-ExpertsRetrieval | CodeCode Available | 0 |
| Scalable Neural Data Server: A Data Recommender for Transfer Learning | Jun 19, 2022 | Mixture-of-ExpertsTransfer Learning | —Unverified | 0 |
| Adaptive Expert Models for Personalization in Federated Learning | Jun 15, 2022 | Federated LearningMixture-of-Experts | CodeCode Available | 0 |
| Towards Universal Sequence Representation Learning for Recommender Systems | Jun 13, 2022 | Mixture-of-ExpertsRecommendation Systems | CodeCode Available | 2 |
| Uni-Perceiver-MoE: Learning Sparse Generalist Models with Conditional MoEs | Jun 9, 2022 | Image CaptioningImage Classification | CodeCode Available | 2 |
| Sparse Mixture-of-Experts are Domain Generalizable Learners | Jun 8, 2022 | Domain GeneralizationMixture-of-Experts | CodeCode Available | 1 |