SOTAVerified

Mixture-of-Experts

Papers

Showing 971980 of 1312 papers

TitleStatusHype
MCR-DL: Mix-and-Match Communication Runtime for Deep Learning0
Scaling Vision-Language Models with Sparse Mixture of Experts0
A Hybrid Tensor-Expert-Data Parallelism Approach to Optimize Mixture-of-Experts Training0
Towards MoE Deployment: Mitigating Inefficiencies in Mixture-of-Expert (MoE) Inference0
Sparse MoE as the New Dropout: Scaling Dense and Self-Slimmable TransformersCode1
MixPHM: Redundancy-Aware Parameter-Efficient Tuning for Low-Resource Visual Question AnsweringCode1
Improving Expert Specialization in Mixture of Experts0
Improved Training of Mixture-of-Experts Language GANs0
TMoE-P: Towards the Pareto Optimum for Multivariate Soft Sensors0
Massively Multilingual Shallow Fusion with Large Language Models0
Show:102550
← PrevPage 98 of 132Next →

No leaderboard results yet.