| FloE: On-the-Fly MoE Inference on Memory-constrained GPU | May 9, 2025 | CPUGPU | —Unverified | 0 |
| fMoE: Fine-Grained Expert Offloading for Large Mixture-of-Experts Serving | Feb 7, 2025 | CPUGPU | —Unverified | 0 |
| FMT:A Multimodal Pneumonia Detection Model Based on Stacking MOE Framework | Mar 7, 2025 | DiagnosticMedical Image Analysis | —Unverified | 0 |
| ForceVLA: Enhancing VLA Models with a Force-aware MoE for Contact-rich Manipulation | May 28, 2025 | Contact-rich ManipulationMixture-of-Experts | —Unverified | 0 |
| Free Agent in Agent-Based Mixture-of-Experts Generative AI Framework | Jan 29, 2025 | Fraud DetectionMixture-of-Experts | —Unverified | 0 |
| FreqMoE: Dynamic Frequency Enhancement for Neural PDE Solvers | May 11, 2025 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| Fresh-CL: Feature Realignment through Experts on Hypersphere in Continual Learning | Jan 4, 2025 | Continual LearningMixture-of-Experts | —Unverified | 0 |
| From Google Gemini to OpenAI Q* (Q-Star): A Survey of Reshaping the Generative Artificial Intelligence (AI) Research Landscape | Dec 18, 2023 | Mixture-of-Experts | —Unverified | 0 |
| FSMoE: A Flexible and Scalable Training System for Sparse Mixture-of-Experts Models | Jan 18, 2025 | GPUMixture-of-Experts | —Unverified | 0 |
| Full-Precision Free Binary Graph Neural Networks | Sep 29, 2021 | Graph Neural NetworkMixture-of-Experts | —Unverified | 0 |