| MoE-Lightning: High-Throughput MoE Inference on Memory-constrained GPUs | Nov 18, 2024 | Computational EfficiencyCPU | —Unverified | 0 |
| Weakly-Supervised Multimodal Learning on MIMIC-CXR | Nov 15, 2024 | Data IntegrationMixture-of-Experts | CodeCode Available | 0 |
| Sparse Upcycling: Inference Inefficient Finetuning | Nov 13, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Lynx: Enabling Efficient MoE Inference through Dynamic Batch-Aware Expert Selection | Nov 13, 2024 | Code GenerationMathematical Reasoning | —Unverified | 0 |
| Towards Vision Mixture of Experts for Wildlife Monitoring on the Edge | Nov 12, 2024 | Mixture-of-Experts | —Unverified | 0 |
| PERFT: Parameter-Efficient Routed Fine-Tuning for Mixture-of-Expert Model | Nov 12, 2024 | Arithmetic ReasoningMixture-of-Experts | —Unverified | 0 |
| Imitation Learning from Observations: An Autoregressive Mixture of Experts Approach | Nov 12, 2024 | Autonomous DrivingImitation Learning | —Unverified | 0 |
| Adaptive Conditional Expert Selection Network for Multi-domain Recommendation | Nov 11, 2024 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| WDMoE: Wireless Distributed Mixture of Experts for Large Language Models | Nov 11, 2024 | Mixture-of-Experts | —Unverified | 0 |
| NeKo: Toward Post Recognition Generative Correction Large Language Models with Task-Oriented Experts | Nov 8, 2024 | Mixture-of-ExpertsOptical Character Recognition (OCR) | —Unverified | 0 |