SOTAVerified

Mixture-of-Experts

Papers

Showing 521530 of 1312 papers

TitleStatusHype
A Mixture of Expert Approach for Low-Cost Customization of Deep Neural Networks0
Hypertext Entity Extraction in Webpage0
HydraSum - Disentangling Stylistic Features in Text Summarization using Multi-Decoder Models0
Hunyuan-TurboS: Advancing Large Language Models through Mamba-Transformer Synergy and Adaptive Chain-of-Thought0
A Universal Approximation Theorem for Mixture of Experts Models0
AMEND: A Mixture of Experts Framework for Long-tailed Trajectory Prediction0
Adaptive Detection of Fast Moving Celestial Objects Using a Mixture of Experts and Physical-Inspired Neural Network0
How to Upscale Neural Networks with Scaling Law? A Survey and Practical Guidelines0
How Lightweight Can A Vision Transformer Be0
How does Architecture Influence the Base Capabilities of Pre-trained Language Models? A Case Study Based on FFN-Wider and MoE Transformers0
Show:102550
← PrevPage 53 of 132Next →

No leaderboard results yet.