| Investigating Mixture of Experts in Dense Retrieval | Dec 16, 2024 | Information RetrievalMixture-of-Experts | —Unverified | 0 |
| Towards Adversarial Robustness of Model-Level Mixture-of-Experts Architectures for Semantic Segmentation | Dec 16, 2024 | Adversarial RobustnessMixture-of-Experts | CodeCode Available | 0 |
| Wonderful Matrices: Combining for a More Efficient and Effective Foundation Model Architecture | Dec 16, 2024 | Mixture-of-ExpertsPosition | CodeCode Available | 1 |
| DeMo: Decoupled Feature-Based Mixture of Experts for Multi-Modal Object Re-Identification | Dec 14, 2024 | Mixture-of-ExpertsObject | CodeCode Available | 2 |
| Llama 3 Meets MoE: Efficient Upcycling | Dec 13, 2024 | Mixture-of-ExpertsMMLU | —Unverified | 0 |
| DeepSeek-VL2: Mixture-of-Experts Vision-Language Models for Advanced Multimodal Understanding | Dec 13, 2024 | Chart UnderstandingMixture-of-Experts | CodeCode Available | 9 |
| Towards a Multimodal Large Language Model with Pixel-Level Insight for Biomedicine | Dec 12, 2024 | Language ModelingLanguage Modelling | CodeCode Available | 2 |
| Mixture of Experts Meets Decoupled Message Passing: Towards General and Adaptive Node Classification | Dec 11, 2024 | Computational Efficiency | CodeCode Available | 0 |
| Adaptive Prompting for Continual Relation Extraction: A Within-Task Variance Perspective | Dec 11, 2024 | Continual Relation ExtractionMixture-of-Experts | —Unverified | 0 |
| MoE-CAP: Benchmarking Cost, Accuracy and Performance of Sparse Mixture-of-Experts Systems | Dec 10, 2024 | BenchmarkingMixture-of-Experts | —Unverified | 0 |