| MM1: Methods, Analysis & Insights from Multimodal LLM Pre-training | Mar 14, 2024 | In-Context LearningMixture-of-Experts | —Unverified | 0 |
| Unleashing the Power of Meta-tuning for Few-shot Generalization Through Sparse Interpolated Experts | Mar 13, 2024 | Domain GeneralizationFew-Shot Image Classification | CodeCode Available | 1 |
| Scattered Mixture-of-Experts Implementation | Mar 13, 2024 | Mixture-of-Experts | CodeCode Available | 2 |
| Conditional computation in neural networks: principles and research trends | Mar 12, 2024 | Mixture-of-Expertsscientific discovery | —Unverified | 0 |
| Harder Tasks Need More Experts: Dynamic Routing in MoE Models | Mar 12, 2024 | Computational EfficiencyMixture-of-Experts | CodeCode Available | 2 |
| Branch-Train-MiX: Mixing Expert LLMs into a Mixture-of-Experts LLM | Mar 12, 2024 | Arithmetic ReasoningCode Generation | —Unverified | 0 |
| Equipping Computational Pathology Systems with Artifact Processing Pipelines: A Showcase for Computation and Performance Trade-offs | Mar 12, 2024 | Airbubbles DetectionAnomaly Detection | CodeCode Available | 0 |
| MoAI: Mixture of All Intelligence for Large Language and Vision Models | Mar 12, 2024 | AllMixture-of-Experts | CodeCode Available | 3 |
| Acquiring Diverse Skills using Curriculum Reinforcement Learning with Mixture of Experts | Mar 11, 2024 | Mixture-of-ExpertsReinforcement Learning (RL) | —Unverified | 0 |
| Unity by Diversity: Improved Representation Learning in Multimodal VAEs | Mar 8, 2024 | DecoderDiversity | CodeCode Available | 1 |