SOTAVerified

Mixture-of-Experts

Papers

Showing 10011050 of 1312 papers

TitleStatusHype
UniCodec: Unified Audio Codec with Single Domain-Adaptive Codebook0
UniCoRN: Latent Diffusion-based Unified Controllable Image Restoration Network across Multiple Degradations0
Unified Modeling of Multi-Domain Multi-Device ASR Systems0
Unify and Anchor: A Context-Aware Transformer for Cross-Domain Time Series Forecasting0
Unimodal-driven Distillation in Multimodal Emotion Recognition with Dynamic Fusion0
UniPaint: Unified Space-time Video Inpainting via Mixture-of-Experts0
UniF^2ace: Fine-grained Face Understanding and Generation with Unified Multimodal Models0
UniUIR: Considering Underwater Image Restoration as An All-in-One Learner0
Unraveling the Localized Latents: Learning Stratified Manifold Structures in LLM Embedding Space with Sparse Mixture-of-Experts0
Unveiling Hidden Collaboration within Mixture-of-Experts in Large Language Models0
UOE: Unlearning One Expert Is Enough For Mixture-of-experts LLMS0
Upcycling Instruction Tuning from Dense to Mixture-of-Experts via Parameter Merging0
Upcycling Large Language Models into Mixture of Experts0
Using Deep Mixture-of-Experts to Detect Word Meaning Shift for TempoWiC0
Utility-Driven Speculative Decoding for Mixture-of-Experts0
Vanilla Transformers are Transfer Capability Teachers0
Variational Distillation of Diffusion Policies into Mixture of Experts0
Variational Mixture of Gaussian Process Experts0
ViMoE: An Empirical Study of Designing Vision Mixture-of-Experts0
Visual Saliency Prediction Using a Mixture of Deep Neural Networks0
WDMoE: Wireless Distributed Large Language Models with Mixture of Experts0
WDMoE: Wireless Distributed Mixture of Experts for Large Language Models0
WeNet: Weighted Networks for Recurrent Network Architecture Search0
Who Says Elephants Can't Run: Bringing Large Scale MoE Models into Cloud Scale Production0
Wolf: Captioning Everything with a World Summarization Framework0
Yi-Lightning Technical Report0
Zero-Resource Multilingual Model Transfer: Learning What to Share0
Multimodal Contrastive Learning with LIMoE: the Language-Image Mixture of Experts0
Multimodal Deep Learning-Empowered Beam Prediction in Future THz ISAC Systems0
Multi-Modal Generative AI: Multi-modal LLM, Diffusion and Beyond0
Multimodal Mixture of Low-Rank Experts for Sentiment Analysis and Emotion Recognition0
Multimodal Variational Autoencoder: a Barycentric View0
Multi-omics data integration for early diagnosis of hepatocellular carcinoma (HCC) using machine learning0
Multi-Task Learning via Task Multi-Clustering0
Multi-Treatment Multi-Task Uplift Modeling for Enhancing User Growth0
Multi-Type Context-Aware Conversational Recommender Systems via Mixture-of-Experts0
Named Entity and Relation Extraction with Multi-Modal Retrieval0
NeKo: Toward Post Recognition Generative Correction Large Language Models with Task-Oriented Experts0
Nested Mixture of Experts: Cooperative and Competitive Learning of Hybrid Dynamical System0
Neural Data Server: A Large-Scale Search Engine for Transfer Learning Data0
Neural Experts: Mixture of Experts for Implicit Neural Representations0
Neural Transduction for Multilingual Lexical Translation0
NeuroMoE: A Transformer-Based Mixture-of-Experts Framework for Multi-Modal Neurological Disorder Classification0
Neutral residues: revisiting adapters for model extension0
NEXT: Multi-Grained Mixture of Experts via Text-Modulation for Multi-Modal Object Re-ID0
Nexus: Specialization meets Adaptability for Efficiently Training Mixture of Experts0
Node-wise Filtering in Graph Neural Networks: A Mixture of Experts Approach0
NoEsis: Differentially Private Knowledge Transfer in Modular LLM Adaptation0
Noise-Robustness Through Noise: Asymmetric LoRA Adaption with Poisoning Expert0
Off-policy Maximum Entropy Reinforcement Learning : Soft Actor-Critic with Advantage Weighted Mixture Policy(SAC-AWMP)0
Show:102550
← PrevPage 21 of 27Next →

No leaderboard results yet.