| Object Detection using Event Camera: A MoE Heat Conduction based Detector and A New Benchmark Dataset | Dec 9, 2024 | Computational EfficiencyMixture-of-Experts | CodeCode Available | 2 |
| UniPaint: Unified Space-time Video Inpainting via Mixture-of-Experts | Dec 9, 2024 | Mixture-of-ExpertsVideo Inpainting | —Unverified | 0 |
| An Entailment Tree Generation Approach for Multimodal Multi-Hop Question Answering with Mixture-of-Experts and Iterative Feedback Mechanism | Dec 8, 2024 | Mixture-of-ExpertsMulti-hop Question Answering | —Unverified | 0 |
| Towards 3D Acceleration for low-power Mixture-of-Experts and Multi-Head Attention Spiking Transformers | Dec 7, 2024 | Mixture-of-Experts | —Unverified | 0 |
| SAME: Learning Generic Language-Guided Visual Navigation with State-Adaptive Mixture of Experts | Dec 7, 2024 | General KnowledgeMixture-of-Experts | CodeCode Available | 1 |
| RSUniVLM: A Unified Vision Language Model for Remote Sensing via Granularity-oriented Mixture of Experts | Dec 7, 2024 | Change DetectionImage Comprehension | CodeCode Available | 1 |
| Steps are all you need: Rethinking STEM Education with Prompt Engineering | Dec 6, 2024 | AllHallucination | —Unverified | 0 |
| Monet: Mixture of Monosemantic Experts for Transformers | Dec 5, 2024 | Dictionary LearningMixture-of-Experts | CodeCode Available | 2 |
| Convolutional Neural Networks and Mixture of Experts for Intrusion Detection in 5G Networks and beyond | Dec 4, 2024 | Intrusion DetectionMixture-of-Experts | —Unverified | 0 |
| Yi-Lightning Technical Report | Dec 2, 2024 | ChatbotLarge Language Model | —Unverified | 0 |
| Mixture of Experts for Node Classification | Nov 30, 2024 | ClassificationMixture-of-Experts | —Unverified | 0 |
| MQFL-FHE: Multimodal Quantum Federated Learning Framework with Fully Homomorphic Encryption | Nov 30, 2024 | Federated LearningMixture-of-Experts | —Unverified | 0 |
| HiMoE: Heterogeneity-Informed Mixture-of-Experts for Fair Spatial-Temporal Forecasting | Nov 30, 2024 | FairnessMixture-of-Experts | —Unverified | 0 |
| LaVIDE: A Language-Vision Discriminator for Detecting Changes in Satellite Image with Map References | Nov 29, 2024 | Change DetectionMixture-of-Experts | —Unverified | 0 |
| On the effectiveness of discrete representations in sparse mixture of experts | Nov 28, 2024 | Mixture-of-ExpertsQuantization | —Unverified | 0 |
| Mixture of Cache-Conditional Experts for Efficient Mobile Device Inference | Nov 27, 2024 | GSM8KLanguage Modeling | —Unverified | 0 |
| Complexity Experts are Task-Discriminative Learners for Any Image Restoration | Nov 27, 2024 | AttributeBlind All-in-One Image Restoration | —Unverified | 0 |
| Mixture of Experts in Image Classification: What's the Sweet Spot? | Nov 27, 2024 | image-classificationImage Classification | —Unverified | 0 |
| UOE: Unlearning One Expert Is Enough For Mixture-of-experts LLMS | Nov 27, 2024 | Large Language ModelMixture-of-Experts | —Unverified | 0 |
| Condense, Don't Just Prune: Enhancing Efficiency and Performance in MoE Layer Pruning | Nov 26, 2024 | Mixture-of-Experts | CodeCode Available | 1 |
| Enhancing Code-Switching ASR Leveraging Non-Peaky CTC Loss and Deep Language Posterior Injection | Nov 26, 2024 | Automatic Speech RecognitionAutomatic Speech Recognition (ASR) | —Unverified | 0 |
| H^3Fusion: Helpful, Harmless, Honest Fusion of Aligned LLMs | Nov 26, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| MH-MoE: Multi-Head Mixture-of-Experts | Nov 25, 2024 | Mixture-of-Experts | —Unverified | 0 |
| LDACP: Long-Delayed Ad Conversions Prediction Model for Bidding Strategy | Nov 25, 2024 | Mixture-of-Expertsregression | —Unverified | 0 |
| LLaMA-MoE v2: Exploring Sparsity of LLaMA from Perspective of Mixture-of-Experts with Post-Training | Nov 24, 2024 | MathMixture-of-Experts | CodeCode Available | 2 |
| Lifelong Knowledge Editing for Vision Language Models with Low-Rank Mixture-of-Experts | Nov 23, 2024 | knowledge editingMixture-of-Experts | —Unverified | 0 |
| MERLOT: A Distilled LLM-based Mixture-of-Experts Framework for Scalable Encrypted Traffic Classification | Nov 20, 2024 | DecoderLanguage Modeling | —Unverified | 0 |
| KAAE: Numerical Reasoning for Knowledge Graphs via Knowledge-aware Attributes Learning | Nov 20, 2024 | AttributeContrastive Learning | —Unverified | 0 |
| Ultra-Sparse Memory Network | Nov 19, 2024 | Mixture-of-Experts | —Unverified | 0 |
| CNMBERT: A Model for Converting Hanyu Pinyin Abbreviations to Chinese Characters | Nov 18, 2024 | fill-maskFill Mask | CodeCode Available | 2 |
| MoE-Lightning: High-Throughput MoE Inference on Memory-constrained GPUs | Nov 18, 2024 | Computational EfficiencyCPU | —Unverified | 0 |
| Awaker2.5-VL: Stably Scaling MLLMs with Parameter-Efficient Mixture of Experts | Nov 16, 2024 | Mixture-of-ExpertsOptical Character Recognition (OCR) | CodeCode Available | 1 |
| Weakly-Supervised Multimodal Learning on MIMIC-CXR | Nov 15, 2024 | Data IntegrationMixture-of-Experts | CodeCode Available | 0 |
| Sparse Upcycling: Inference Inefficient Finetuning | Nov 13, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Lynx: Enabling Efficient MoE Inference through Dynamic Batch-Aware Expert Selection | Nov 13, 2024 | Code GenerationMathematical Reasoning | —Unverified | 0 |
| Imitation Learning from Observations: An Autoregressive Mixture of Experts Approach | Nov 12, 2024 | Autonomous DrivingImitation Learning | —Unverified | 0 |
| PERFT: Parameter-Efficient Routed Fine-Tuning for Mixture-of-Expert Model | Nov 12, 2024 | Arithmetic ReasoningMixture-of-Experts | —Unverified | 0 |
| Towards Vision Mixture of Experts for Wildlife Monitoring on the Edge | Nov 12, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Adaptive Conditional Expert Selection Network for Multi-domain Recommendation | Nov 11, 2024 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| WDMoE: Wireless Distributed Mixture of Experts for Large Language Models | Nov 11, 2024 | Mixture-of-Experts | —Unverified | 0 |
| NeKo: Toward Post Recognition Generative Correction Large Language Models with Task-Oriented Experts | Nov 8, 2024 | Mixture-of-ExpertsOptical Character Recognition (OCR) | —Unverified | 0 |
| DA-MoE: Addressing Depth-Sensitivity in Graph-Level Analysis through Mixture of Experts | Nov 5, 2024 | Mixture-of-ExpertsSensitivity | CodeCode Available | 0 |
| Advancing Robust Underwater Acoustic Target Recognition through Multi-task Learning and Multi-Gate Mixture-of-Experts | Nov 5, 2024 | Mixture-of-ExpertsMulti-Task Learning | —Unverified | 0 |
| FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation | Nov 4, 2024 | Federated LearningMixture-of-Experts | —Unverified | 0 |
| Hunyuan-Large: An Open-Source MoE Model with 52 Billion Activated Parameters by Tencent | Nov 4, 2024 | Logical ReasoningMathematical Problem-Solving | CodeCode Available | 5 |
| RS-MoE: Mixture of Experts for Remote Sensing Image Captioning and Visual Question Answering | Nov 3, 2024 | DescriptiveImage Captioning | —Unverified | 0 |
| HOBBIT: A Mixed Precision Expert Offloading System for Fast MoE Inference | Nov 3, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Facet-Aware Multi-Head Mixture-of-Experts Model for Sequential Recommendation | Nov 3, 2024 | Mixture-of-ExpertsSequential Recommendation | —Unverified | 0 |
| PMoL: Parameter Efficient MoE for Preference Mixing of LLM Alignment | Nov 2, 2024 | Mixture-of-Experts | —Unverified | 0 |
| SLED: Self Logits Evolution Decoding for Improving Factuality in Large Language Models | Nov 1, 2024 | Mixture-of-Experts | CodeCode Available | 2 |