| Towards Adversarial Robustness of Model-Level Mixture-of-Experts Architectures for Semantic Segmentation | Dec 16, 2024 | Adversarial RobustnessMixture-of-Experts | CodeCode Available | 0 |
| Enhancing Healthcare Recommendation Systems with a Multimodal LLMs-based MOE Architecture | Dec 16, 2024 | Mixture-of-ExpertsRecommendation Systems | —Unverified | 0 |
| Investigating Mixture of Experts in Dense Retrieval | Dec 16, 2024 | Information RetrievalMixture-of-Experts | —Unverified | 0 |
| Llama 3 Meets MoE: Efficient Upcycling | Dec 13, 2024 | Mixture-of-ExpertsMMLU | —Unverified | 0 |
| Mixture of Experts Meets Decoupled Message Passing: Towards General and Adaptive Node Classification | Dec 11, 2024 | Computational Efficiency | CodeCode Available | 0 |
| Adaptive Prompting for Continual Relation Extraction: A Within-Task Variance Perspective | Dec 11, 2024 | Continual Relation ExtractionMixture-of-Experts | —Unverified | 0 |
| MoE-CAP: Benchmarking Cost, Accuracy and Performance of Sparse Mixture-of-Experts Systems | Dec 10, 2024 | BenchmarkingMixture-of-Experts | —Unverified | 0 |
| UniPaint: Unified Space-time Video Inpainting via Mixture-of-Experts | Dec 9, 2024 | Mixture-of-ExpertsVideo Inpainting | —Unverified | 0 |
| An Entailment Tree Generation Approach for Multimodal Multi-Hop Question Answering with Mixture-of-Experts and Iterative Feedback Mechanism | Dec 8, 2024 | Mixture-of-ExpertsMulti-hop Question Answering | —Unverified | 0 |
| Towards 3D Acceleration for low-power Mixture-of-Experts and Multi-Head Attention Spiking Transformers | Dec 7, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Steps are all you need: Rethinking STEM Education with Prompt Engineering | Dec 6, 2024 | AllHallucination | —Unverified | 0 |
| Convolutional Neural Networks and Mixture of Experts for Intrusion Detection in 5G Networks and beyond | Dec 4, 2024 | Intrusion DetectionMixture-of-Experts | —Unverified | 0 |
| Yi-Lightning Technical Report | Dec 2, 2024 | ChatbotLarge Language Model | —Unverified | 0 |
| HiMoE: Heterogeneity-Informed Mixture-of-Experts for Fair Spatial-Temporal Forecasting | Nov 30, 2024 | FairnessMixture-of-Experts | —Unverified | 0 |
| MQFL-FHE: Multimodal Quantum Federated Learning Framework with Fully Homomorphic Encryption | Nov 30, 2024 | Federated LearningMixture-of-Experts | —Unverified | 0 |
| Mixture of Experts for Node Classification | Nov 30, 2024 | ClassificationMixture-of-Experts | —Unverified | 0 |
| LaVIDE: A Language-Vision Discriminator for Detecting Changes in Satellite Image with Map References | Nov 29, 2024 | Change DetectionMixture-of-Experts | —Unverified | 0 |
| On the effectiveness of discrete representations in sparse mixture of experts | Nov 28, 2024 | Mixture-of-ExpertsQuantization | —Unverified | 0 |
| Mixture of Cache-Conditional Experts for Efficient Mobile Device Inference | Nov 27, 2024 | GSM8KLanguage Modeling | —Unverified | 0 |
| UOE: Unlearning One Expert Is Enough For Mixture-of-experts LLMS | Nov 27, 2024 | Large Language ModelMixture-of-Experts | —Unverified | 0 |
| Mixture of Experts in Image Classification: What's the Sweet Spot? | Nov 27, 2024 | image-classificationImage Classification | —Unverified | 0 |
| Complexity Experts are Task-Discriminative Learners for Any Image Restoration | Nov 27, 2024 | AttributeBlind All-in-One Image Restoration | —Unverified | 0 |
| Enhancing Code-Switching ASR Leveraging Non-Peaky CTC Loss and Deep Language Posterior Injection | Nov 26, 2024 | Automatic Speech RecognitionAutomatic Speech Recognition (ASR) | —Unverified | 0 |
| H^3Fusion: Helpful, Harmless, Honest Fusion of Aligned LLMs | Nov 26, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| LDACP: Long-Delayed Ad Conversions Prediction Model for Bidding Strategy | Nov 25, 2024 | Mixture-of-Expertsregression | —Unverified | 0 |
| MH-MoE: Multi-Head Mixture-of-Experts | Nov 25, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Lifelong Knowledge Editing for Vision Language Models with Low-Rank Mixture-of-Experts | Nov 23, 2024 | knowledge editingMixture-of-Experts | —Unverified | 0 |
| MERLOT: A Distilled LLM-based Mixture-of-Experts Framework for Scalable Encrypted Traffic Classification | Nov 20, 2024 | DecoderLanguage Modeling | —Unverified | 0 |
| KAAE: Numerical Reasoning for Knowledge Graphs via Knowledge-aware Attributes Learning | Nov 20, 2024 | AttributeContrastive Learning | —Unverified | 0 |
| Ultra-Sparse Memory Network | Nov 19, 2024 | Mixture-of-Experts | —Unverified | 0 |
| MoE-Lightning: High-Throughput MoE Inference on Memory-constrained GPUs | Nov 18, 2024 | Computational EfficiencyCPU | —Unverified | 0 |
| Weakly-Supervised Multimodal Learning on MIMIC-CXR | Nov 15, 2024 | Data IntegrationMixture-of-Experts | CodeCode Available | 0 |
| Sparse Upcycling: Inference Inefficient Finetuning | Nov 13, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Lynx: Enabling Efficient MoE Inference through Dynamic Batch-Aware Expert Selection | Nov 13, 2024 | Code GenerationMathematical Reasoning | —Unverified | 0 |
| Towards Vision Mixture of Experts for Wildlife Monitoring on the Edge | Nov 12, 2024 | Mixture-of-Experts | —Unverified | 0 |
| PERFT: Parameter-Efficient Routed Fine-Tuning for Mixture-of-Expert Model | Nov 12, 2024 | Arithmetic ReasoningMixture-of-Experts | —Unverified | 0 |
| Imitation Learning from Observations: An Autoregressive Mixture of Experts Approach | Nov 12, 2024 | Autonomous DrivingImitation Learning | —Unverified | 0 |
| Adaptive Conditional Expert Selection Network for Multi-domain Recommendation | Nov 11, 2024 | Computational EfficiencyMixture-of-Experts | —Unverified | 0 |
| WDMoE: Wireless Distributed Mixture of Experts for Large Language Models | Nov 11, 2024 | Mixture-of-Experts | —Unverified | 0 |
| NeKo: Toward Post Recognition Generative Correction Large Language Models with Task-Oriented Experts | Nov 8, 2024 | Mixture-of-ExpertsOptical Character Recognition (OCR) | —Unverified | 0 |
| DA-MoE: Addressing Depth-Sensitivity in Graph-Level Analysis through Mixture of Experts | Nov 5, 2024 | Mixture-of-ExpertsSensitivity | CodeCode Available | 0 |
| Advancing Robust Underwater Acoustic Target Recognition through Multi-task Learning and Multi-Gate Mixture-of-Experts | Nov 5, 2024 | Mixture-of-ExpertsMulti-Task Learning | —Unverified | 0 |
| FedMoE-DA: Federated Mixture of Experts via Domain Aware Fine-grained Aggregation | Nov 4, 2024 | Federated LearningMixture-of-Experts | —Unverified | 0 |
| HOBBIT: A Mixed Precision Expert Offloading System for Fast MoE Inference | Nov 3, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Facet-Aware Multi-Head Mixture-of-Experts Model for Sequential Recommendation | Nov 3, 2024 | Mixture-of-ExpertsSequential Recommendation | —Unverified | 0 |
| RS-MoE: Mixture of Experts for Remote Sensing Image Captioning and Visual Question Answering | Nov 3, 2024 | DescriptiveImage Captioning | —Unverified | 0 |
| PMoL: Parameter Efficient MoE for Preference Mixing of LLM Alignment | Nov 2, 2024 | Mixture-of-Experts | —Unverified | 0 |
| MoNTA: Accelerating Mixture-of-Experts Training with Network-Traffc-Aware Parallel Optimization | Nov 1, 2024 | 8kMixture-of-Experts | CodeCode Available | 0 |
| MoE-I^2: Compressing Mixture of Experts Models through Inter-Expert Pruning and Intra-Expert Low-Rank Decomposition | Nov 1, 2024 | Mixture-of-Experts | CodeCode Available | 0 |
| Stereo-Talker: Audio-driven 3D Human Synthesis with Prior-Guided Mixture-of-Experts | Oct 31, 2024 | Language ModelingLanguage Modelling | —Unverified | 0 |