| NeKo: Toward Post Recognition Generative Correction Large Language Models with Task-Oriented Experts | Nov 8, 2024 | Mixture-of-ExpertsOptical Character Recognition (OCR) | —Unverified | 0 |
| DA-MoE: Addressing Depth-Sensitivity in Graph-Level Analysis through Mixture of Experts | Nov 5, 2024 | Mixture-of-ExpertsSensitivity | CodeCode Available | 0 |
| Advancing Robust Underwater Acoustic Target Recognition through Multi-task Learning and Multi-Gate Mixture-of-Experts | Nov 5, 2024 | Mixture-of-ExpertsMulti-Task Learning | —Unverified | 0 |
| FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation | Nov 4, 2024 | Federated LearningMixture-of-Experts | —Unverified | 0 |
| Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent | Nov 4, 2024 | Logical ReasoningMathematical Problem-Solving | CodeCode Available | 5 |
| RS-MoE: Mixture of Experts for Remote Sensing Image Captioning and Visual Question Answering | Nov 3, 2024 | DescriptiveImage Captioning | —Unverified | 0 |
| HOBBIT: A Mixed Precision Expert Offloading System for Fast MoE Inference | Nov 3, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Facet-Aware Multi-Head Mixture-of-Experts Model for Sequential Recommendation | Nov 3, 2024 | Mixture-of-ExpertsSequential Recommendation | —Unverified | 0 |
| PMoL: Parameter Efficient MoE for Preference Mixing of LLM Alignment | Nov 2, 2024 | Mixture-of-Experts | —Unverified | 0 |
| SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language Models | Nov 1, 2024 | Mixture-of-Experts | CodeCode Available | 2 |