| Theory on Mixture-of-Experts in Continual Learning | Jun 24, 2024 | Continual LearningMixture-of-Experts | —Unverified | 0 |
| The power of fine-grained experts: Granularity boosts expressivity in Mixture of Experts | May 11, 2025 | Mixture-of-Experts | —Unverified | 0 |
| The Ultimate Guide to Fine-Tuning LLMs from Basics to Breakthroughs: An Exhaustive Review of Technologies, Research, Best Practices, Applied Research Challenges and Opportunities | Aug 23, 2024 | Computational EfficiencyInference Optimization | —Unverified | 0 |
| THOR-MoE: Hierarchical Task-Guided and Context-Responsive Routing for Neural Machine Translation | May 20, 2025 | Machine TranslationMixture-of-Experts | —Unverified | 0 |
| Time series forecasting with high stakes: A field study of the air cargo industry | Jul 29, 2024 | Decision MakingDemand Forecasting | —Unverified | 0 |
| Time Tracker: Mixture-of-Experts-Enhanced Foundation Time Series Forecasting Model with Decoupled Training Pipelines | May 21, 2025 | Graph LearningMixture-of-Experts | —Unverified | 0 |
| Tiny-Attention Adapter: Contexts Are More Important Than the Number of Parameters | Oct 18, 2022 | Language ModelingLanguage Modelling | —Unverified | 0 |
| TMoE-P: Towards the Pareto Optimum for Multivariate Soft Sensors | Feb 21, 2023 | Mixture-of-Experts | —Unverified | 0 |
| ToMoE: Converting Dense Large Language Models to Mixture-of-Experts through Dynamic Structural Pruning | Jan 25, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Topic Compositional Neural Language Model | Dec 28, 2017 | Language ModelingLanguage Modelling | —Unverified | 0 |
| To Repeat or Not To Repeat: Insights from Scaling LLM under Token-Crisis | May 22, 2023 | Mixture-of-Experts | —Unverified | 0 |
| Toward Mixture-of-Experts Enabled Trustworthy Semantic Communication for 6G Networks | Sep 24, 2024 | Mixture-of-ExpertsSemantic Communication | —Unverified | 0 |
| Towards 3D Acceleration for low-power Mixture-of-Experts and Multi-Head Attention Spiking Transformers | Dec 7, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Towards A Better Metric for Text-to-Video Generation | Jan 15, 2024 | Mixture-of-ExpertsText-to-Video Generation | —Unverified | 0 |
| Towards an empirical understanding of MoE design choices | Feb 20, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Towards A Unified View of Sparse Feed-Forward Network in Pretraining Large Language Model | May 23, 2023 | AvgLanguage Modeling | —Unverified | 0 |
| Towards Convergence Rates for Parameter Estimation in Gaussian-gated Mixture of Experts | May 12, 2023 | Ensemble LearningMixture-of-Experts | —Unverified | 0 |
| Towards Efficient Foundation Model for Zero-shot Amodal Segmentation | Jan 1, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Towards Efficient Single Image Dehazing and Desnowing | Apr 19, 2022 | Image DehazingImage Restoration | —Unverified | 0 |
| Towards Foundational Models for Dynamical System Reconstruction: Hierarchical Meta-Learning via Mixture of Experts | Feb 7, 2025 | Meta-LearningMixture-of-Experts | —Unverified | 0 |
| Towards Lightweight Neural Animation : Exploration of Neural Network Pruning in Mixture of Experts-based Animation Models | Jan 11, 2022 | Mixture-of-ExpertsNetwork Pruning | —Unverified | 0 |
| Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference | Mar 10, 2023 | CPUDecoder | —Unverified | 0 |
| Towards Personalized Federated Multi-Scenario Multi-Task Recommendation | Jun 27, 2024 | Federated LearningMixture-of-Experts | —Unverified | 0 |
| Towards Smart Point-and-Shoot Photography | May 6, 2025 | Mixture-of-ExpertsWord Embeddings | —Unverified | 0 |
| Towards Vision Mixture of Experts for Wildlife Monitoring on the Edge | Nov 12, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Training-efficient density quantum machine learning | May 30, 2024 | LEMMAMixture-of-Experts | —Unverified | 0 |
| Training of Neural Networks with Uncertain Data: A Mixture of Experts Approach | Dec 13, 2023 | Autonomous DrivingMixture-of-Experts | —Unverified | 0 |
| TrajMoE: Spatially-Aware Mixture of Experts for Unified Human Mobility Modeling | May 24, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Transformer Layer Injection: A Novel Approach for Efficient Upscaling of Large Language Models | Oct 15, 2024 | Mixture-of-Experts | —Unverified | 0 |
| Tree-gated Deep Mixture-of-Experts For Pose-robust Face Alignment | Oct 21, 2019 | Face AlignmentMixture-of-Experts | —Unverified | 0 |
| Trend Filtered Mixture of Experts for Automated Gating of High-Frequency Flow Cytometry Data | Apr 16, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Towards Incremental Learning in Large Language Models: A Critical Review | Apr 28, 2024 | Continual LearningIncremental Learning | —Unverified | 0 |
| True Zero-Shot Inference of Dynamical Systems Preserving Long-Term Statistics | May 19, 2025 | Mixture-of-ExpertsTime Series | —Unverified | 0 |
| TS-RAG: Retrieval-Augmented Generation based Time Series Foundation Models are Stronger Zero-Shot Forecaster | Mar 6, 2025 | Domain AdaptationMixture-of-Experts | —Unverified | 0 |
| Tuning of Mixture-of-Experts Mixed-Precision Neural Networks | Sep 29, 2022 | image-classificationImage Classification | —Unverified | 0 |
| Turn Waste into Worth: Rectifying Top-k Router of MoE | Feb 17, 2024 | Computational EfficiencyGPU | —Unverified | 0 |
| Two Experts Are All You Need for Steering Thinking: Reinforcing Cognitive Effort in MoE Reasoning Models Without Additional Training | May 20, 2025 | AllDomain Generalization | —Unverified | 0 |
| Two Is Better Than One: Rotations Scale LoRAs | May 29, 2025 | Mixture-of-Experts | —Unverified | 0 |
| U2++ MoE: Scaling 4.7x parameters with minimal impact on RTF | Apr 25, 2024 | Automatic Speech RecognitionAutomatic Speech Recognition (ASR) | —Unverified | 0 |
| UGG-ReID: Uncertainty-Guided Graph Model for Multi-Modal Object Re-Identification | Jul 7, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Fast Deep Mixtures of Gaussian Process Experts | Jun 11, 2020 | Gaussian ProcessesMixture-of-Experts | —Unverified | 0 |
| Ultra-Sparse Memory Network | Nov 19, 2024 | Mixture-of-Experts | —Unverified | 0 |
| UME: Upcycling Mixture-of-Experts for Scalable and Efficient Automatic Speech Recognition | Dec 23, 2024 | Automatic Speech RecognitionAutomatic Speech Recognition (ASR) | —Unverified | 0 |
| UMoE: Unifying Attention and FFN with Shared Experts | May 12, 2025 | Mixture-of-Experts | —Unverified | 0 |
| Unbiased Gradient Estimation with Balanced Assignments for Mixtures of Experts | Sep 24, 2021 | Mixture-of-Experts | —Unverified | 0 |
| Uncertainty-Aware Driver Trajectory Prediction at Urban Intersections | Jan 16, 2019 | Mixture-of-ExpertsPrediction | —Unverified | 0 |
| Uncertainty-Encoded Multi-Modal Fusion for Robust Object Detection in Autonomous Driving | Jul 30, 2023 | Autonomous DrivingMixture-of-Experts | —Unverified | 0 |
| Understanding Expert Structures on Minimax Parameter Estimation in Contaminated Mixture of Experts | Oct 16, 2024 | Mixture-of-Expertsparameter estimation | —Unverified | 0 |
| UniAdapt: A Universal Adapter for Knowledge Calibration | Oct 1, 2024 | Mixture-of-ExpertsModel Editing | —Unverified | 0 |
| UNIALIGN: Scaling Multimodal Alignment within One Unified Model | Jan 1, 2025 | Mixture-of-Experts | —Unverified | 0 |