| A Survey on Model MoErging: Recycling and Routing Among Specialized Experts for Collaborative Learning | Aug 13, 2024 | Mixture-of-ExpertsSurvey | —Unverified | 0 |
| HoME: Hierarchy of Multi-Gate Experts for Multi-Task Learning at Kuaishou | Aug 10, 2024 | Mixture-of-ExpertsMulti-Task Learning | —Unverified | 0 |
| LaDiMo: Layer-wise Distillation Inspired MoEfier | Aug 8, 2024 | Knowledge DistillationMixture-of-Experts | —Unverified | 0 |
| Understanding the Performance and Estimating the Cost of LLM Fine-Tuning | Aug 8, 2024 | GPUMixture-of-Experts | CodeCode Available | 0 |
| MoC-System: Efficient Fault Tolerance for Sparse Mixture-of-Experts Model Training | Aug 8, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Mixture-of-Noises Enhanced Forgery-Aware Predictor for Multi-Face Manipulation Detection and Localization | Aug 5, 2024 | Face DetectionMixture-of-Experts | —Unverified | 0 |
| HMDN: Hierarchical Multi-Distribution Network for Click-Through Rate Prediction | Aug 2, 2024 | Click-Through Rate PredictionMixture-of-Experts | —Unverified | 0 |
| Multimodal Fusion and Coherence Modeling for Video Topic Segmentation | Aug 1, 2024 | Contrastive LearningMixture-of-Experts | —Unverified | 0 |
| MoMa: Efficient Early-Fusion Pre-training with Mixture of Modality-Aware Experts | Jul 31, 2024 | Causal InferenceLanguage Modelling | —Unverified | 0 |
| PMoE: Progressive Mixture of Experts with Asymmetric Transformer for Continual Learning | Jul 31, 2024 | Continual LearningGeneral Knowledge | —Unverified | 0 |
| Distribution Learning for Molecular Regression | Jul 30, 2024 | Mixture-of-ExpertsMolecular Property Prediction | —Unverified | 0 |
| Time series forecasting with high stakes: A field study of the air cargo industry | Jul 29, 2024 | Decision MakingDemand Forecasting | —Unverified | 0 |
| Mixture of Nested Experts: Adaptive Processing of Visual Tokens | Jul 29, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| Mixture of Modular Experts: Distilling Knowledge from a Multilingual Teacher into Specialized Modular Language Models | Jul 28, 2024 | Knowledge DistillationMixture-of-Experts | CodeCode Available | 0 |
| MOoSE: Multi-Orientation Sharing Experts for Open-set Scene Text Recognition | Jul 26, 2024 | Mixture-of-ExpertsScene Text Recognition | CodeCode Available | 0 |
| Wolf: Captioning Everything with a World Summarization Framework | Jul 26, 2024 | Autonomous DrivingMixture-of-Experts | —Unverified | 0 |
| How Lightweight Can A Vision Transformer Be | Jul 25, 2024 | Mixture-of-ExpertsTransfer Learning | —Unverified | 0 |
| Exploring Domain Robust Lightweight Reward Models based on Router Mechanism | Jul 24, 2024 | Language ModelingLanguage Modelling | —Unverified | 0 |
| Wonderful Matrices: More Efficient and Effective Architecture for Language Modeling Tasks | Jul 24, 2024 | Language ModelingLanguage Modelling | —Unverified | 0 |
| EEGMamba: Bidirectional State Space Model with Mixture of Experts for EEG Multi-task Classification | Jul 20, 2024 | EEGElectroencephalogram (EEG) | —Unverified | 0 |
| EVLM: An Efficient Vision-Language Model for Visual Understanding | Jul 19, 2024 | Image CaptioningLanguage Modeling | —Unverified | 0 |
| Mixture of Experts with Mixture of Precisions for Tuning Quality of Service | Jul 19, 2024 | CPUGPU | —Unverified | 0 |
| Mixture of Experts based Multi-task Supervise Learning from Crowds | Jul 18, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Discussion: Effective and Interpretable Outcome Prediction by Training Sparse Mixtures of Linear Experts | Jul 18, 2024 | feature selectionMixture-of-Experts | —Unverified | 0 |
| MoE-DiffIR: Task-customized Diffusion Priors for Universal Compressed Image Restoration | Jul 15, 2024 | Image RestorationMixture-of-Experts | —Unverified | 0 |
| Boost Your NeRF: A Model-Agnostic Mixture of Experts Framework for High Quality and Efficient Rendering | Jul 15, 2024 | Mixture-of-ExpertsNeRF | —Unverified | 0 |
| MaskMoE: Boosting Token-Level Learning via Routing Mask in Mixture-of-Experts | Jul 13, 2024 | DiversityMixture-of-Experts | CodeCode Available | 0 |
| Diversifying the Expert Knowledge for Task-Agnostic Pruning in Sparse Mixture-of-Experts | Jul 12, 2024 | Mixture-of-Experts | —Unverified | 0 |
| An Unsupervised Domain Adaptation Method for Locating Manipulated Region in partially fake Audio | Jul 11, 2024 | Data AugmentationDiversity | —Unverified | 0 |
| MoVEInt: Mixture of Variational Experts for Learning Human-Robot Interactions from Demonstrations | Jul 10, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| A Simple Architecture for Enterprise Large Language Model Applications based on Role based security and Clearance Levels using Retrieval-Augmented Generation or Mixture of Experts | Jul 9, 2024 | Language ModelingLanguage Modelling | —Unverified | 0 |
| SAM-Med3D-MoE: Towards a Non-Forgetting Segment Anything Model via Mixture of Experts for 3D Medical Image Segmentation | Jul 6, 2024 | General KnowledgeImage Segmentation | —Unverified | 0 |
| Completed Feature Disentanglement Learning for Multimodal MRIs Analysis | Jul 6, 2024 | DisentanglementMixture-of-Experts | CodeCode Available | 0 |
| MobileFlow: A Multimodal LLM For Mobile GUI Agent | Jul 5, 2024 | Action AnalysisLanguage Modelling | —Unverified | 0 |
| Lazarus: Resilient and Elastic Training of Mixture-of-Experts Models with Adaptive Expert Placement | Jul 5, 2024 | GPUMixture-of-Experts | —Unverified | 0 |
| Terminating Differentiable Tree Experts | Jul 2, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Investigating the potential of Sparse Mixtures-of-Experts for multi-domain neural machine translation | Jul 1, 2024 | Machine TranslationMixture-of-Experts | —Unverified | 0 |
| Sparse Diffusion Policy: A Sparse, Reusable, and Flexible Policy for Robot Learning | Jul 1, 2024 | Continual LearningMixture-of-Experts | —Unverified | 0 |
| LEMoE: Advanced Mixture of Experts Adaptor for Lifelong Model Editing of Large Language Models | Jun 28, 2024 | Mixture-of-ExpertsModel Editing | —Unverified | 0 |
| A Teacher Is Worth A Million Instructions | Jun 27, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| Towards Personalized Federated Multi-Scenario Multi-Task Recommendation | Jun 27, 2024 | Federated LearningMixture-of-Experts | —Unverified | 0 |
| SC-MoE: Switch Conformer Mixture of Experts for Unified Streaming and Non-streaming Code-Switching ASR | Jun 26, 2024 | Automatic Speech RecognitionAutomatic Speech Recognition (ASR) | —Unverified | 0 |
| Mixture of Experts in a Mixture of RL settings | Jun 26, 2024 | Deep Reinforcement LearningMixture-of-Experts | —Unverified | 0 |
| MoESD: Mixture of Experts Stable Diffusion to Mitigate Gender Bias | Jun 25, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Peirce in the Machine: How Mixture of Experts Models Perform Hypothesis Construction | Jun 24, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| OTCE: Hybrid SSM and Attention with Cross Domain Mixture of Experts to construct Observer-Thinker-Conceiver-Expresser | Jun 24, 2024 | Language ModelingLanguage Modelling | CodeCode Available | 0 |
| Theory on Mixture-of-Experts in Continual Learning | Jun 24, 2024 | Continual LearningMixture-of-Experts | —Unverified | 0 |
| SimSMoE: Solving Representational Collapse via Similarity Measure | Jun 22, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Low-Rank Mixture-of-Experts for Continual Medical Image Segmentation | Jun 19, 2024 | Continual LearningImage Segmentation | —Unverified | 0 |
| P-Tailor: Customizing Personality Traits for Language Models via Mixture of Specialized LoRA Experts | Jun 18, 2024 | Mixture-of-Experts | —Unverified | 0 |