SOTAVerified

Mixture-of-Experts

Papers

Showing 9511000 of 1312 papers

TitleStatusHype
Theory on Mixture-of-Experts in Continual Learning0
The power of fine-grained experts: Granularity boosts expressivity in Mixture of Experts0
The Ultimate Guide to Fine-Tuning LLMs from Basics to Breakthroughs: An Exhaustive Review of Technologies, Research, Best Practices, Applied Research Challenges and Opportunities0
THOR-MoE: Hierarchical Task-Guided and Context-Responsive Routing for Neural Machine Translation0
Time series forecasting with high stakes: A field study of the air cargo industry0
Time Tracker: Mixture-of-Experts-Enhanced Foundation Time Series Forecasting Model with Decoupled Training Pipelines0
Tiny-Attention Adapter: Contexts Are More Important Than the Number of Parameters0
TMoE-P: Towards the Pareto Optimum for Multivariate Soft Sensors0
ToMoE: Converting Dense Large Language Models to Mixture-of-Experts through Dynamic Structural Pruning0
Topic Compositional Neural Language Model0
To Repeat or Not To Repeat: Insights from Scaling LLM under Token-Crisis0
Toward Mixture-of-Experts Enabled Trustworthy Semantic Communication for 6G Networks0
Towards 3D Acceleration for low-power Mixture-of-Experts and Multi-Head Attention Spiking Transformers0
Towards A Better Metric for Text-to-Video Generation0
Towards an empirical understanding of MoE design choices0
Towards A Unified View of Sparse Feed-Forward Network in Pretraining Large Language Model0
Towards Convergence Rates for Parameter Estimation in Gaussian-gated Mixture of Experts0
Towards Efficient Foundation Model for Zero-shot Amodal Segmentation0
Towards Efficient Single Image Dehazing and Desnowing0
Towards Foundational Models for Dynamical System Reconstruction: Hierarchical Meta-Learning via Mixture of Experts0
Towards Lightweight Neural Animation : Exploration of Neural Network Pruning in Mixture of Experts-based Animation Models0
Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference0
Towards Personalized Federated Multi-Scenario Multi-Task Recommendation0
Towards Smart Point-and-Shoot Photography0
Towards Vision Mixture of Experts for Wildlife Monitoring on the Edge0
Training-efficient density quantum machine learning0
Training of Neural Networks with Uncertain Data: A Mixture of Experts Approach0
TrajMoE: Spatially-Aware Mixture of Experts for Unified Human Mobility Modeling0
Transformer Layer Injection: A Novel Approach for Efficient Upscaling of Large Language Models0
Tree-gated Deep Mixture-of-Experts For Pose-robust Face Alignment0
Trend Filtered Mixture of Experts for Automated Gating of High-Frequency Flow Cytometry Data0
Towards Incremental Learning in Large Language Models: A Critical Review0
True Zero-Shot Inference of Dynamical Systems Preserving Long-Term Statistics0
TS-RAG: Retrieval-Augmented Generation based Time Series Foundation Models are Stronger Zero-Shot Forecaster0
Tuning of Mixture-of-Experts Mixed-Precision Neural Networks0
Turn Waste into Worth: Rectifying Top-k Router of MoE0
Two Experts Are All You Need for Steering Thinking: Reinforcing Cognitive Effort in MoE Reasoning Models Without Additional Training0
Two Is Better Than One: Rotations Scale LoRAs0
U2++ MoE: Scaling 4.7x parameters with minimal impact on RTF0
UGG-ReID: Uncertainty-Guided Graph Model for Multi-Modal Object Re-Identification0
Fast Deep Mixtures of Gaussian Process Experts0
Ultra-Sparse Memory Network0
UME: Upcycling Mixture-of-Experts for Scalable and Efficient Automatic Speech Recognition0
UMoE: Unifying Attention and FFN with Shared Experts0
Unbiased Gradient Estimation with Balanced Assignments for Mixtures of Experts0
Uncertainty-Aware Driver Trajectory Prediction at Urban Intersections0
Uncertainty-Encoded Multi-Modal Fusion for Robust Object Detection in Autonomous Driving0
Understanding Expert Structures on Minimax Parameter Estimation in Contaminated Mixture of Experts0
UniAdapt: A Universal Adapter for Knowledge Calibration0
UNIALIGN: Scaling Multimodal Alignment within One Unified Model0
Show:102550
← PrevPage 20 of 27Next →

No leaderboard results yet.