| Contextual Mixture of Experts: Integrating Knowledge into Predictive Modeling | Nov 1, 2022 | Mixture-of-Experts | —Unverified | 0 |
| Prediction Sets for High-Dimensional Mixture of Experts Models | Oct 30, 2022 | Mixture-of-ExpertsPrediction | —Unverified | 0 |
| Knowledge-in-Context: Towards Knowledgeable Semi-Parametric Language Models | Oct 28, 2022 | Common Sense ReasoningCoreference Resolution | —Unverified | 0 |
| Coordination with Humans via Strategy Matching | Oct 27, 2022 | Mixture-of-Experts | —Unverified | 0 |
| M^3ViT: Mixture-of-Experts Vision Transformer for Efficient Multi-task Learning with Model-Accelerator Co-design | Oct 26, 2022 | Mixture-of-ExpertsMulti-Task Learning | CodeCode Available | 1 |
| On the Adversarial Robustness of Mixture of Experts | Oct 19, 2022 | Adversarial RobustnessMixture-of-Experts | —Unverified | 0 |
| Tiny-Attention Adapter: Contexts Are More Important Than the Number of Parameters | Oct 18, 2022 | Language ModelingLanguage Modelling | —Unverified | 0 |
| AutoMoE: Heterogeneous Mixture-of-Experts with Adaptive Computation for Efficient Neural Machine Translation | Oct 14, 2022 | CPUMachine Translation | CodeCode Available | 1 |
| Mixture of Attention Heads: Selecting Attention Heads Per Token | Oct 11, 2022 | Computational EfficiencyLanguage Modeling | CodeCode Available | 1 |
| FEAMOE: Fair, Explainable and Adaptive Mixture of Experts | Oct 10, 2022 | FairnessMixture-of-Experts | —Unverified | 0 |