SOTAVerified

Mixture-of-Experts

Papers

Showing 10511100 of 1312 papers

TitleStatusHype
Omni-SMoLA: Boosting Generalist Multimodal Models with Soft Mixture of Low-rank Experts0
OMoE: Diversifying Mixture of Low-Rank Adaptation by Orthogonal Finetuning0
On component interactions in two-stage recommender systems0
SkillNet-NLU: A Sparsely Activated Model for General-Purpose Natural Language Understanding0
OneRec: Unifying Retrieve and Rank with Generative Recommender and Iterative Preference Alignment0
One Student Knows All Experts Know: From Sparse to Dense0
On Expert Estimation in Hierarchical Mixture of Experts: Beyond Softmax Gating Functions0
On Least Square Estimation in Softmax Gating Mixture of Experts0
On Minimax Estimation of Parameters in Softmax-Contaminated Mixture of Experts0
On Multi-objective Policy Optimization as a Tool for Reinforcement Learning: Case Studies in Offline RL and Finetuning0
On Parameter Estimation in Deviated Gaussian Mixture of Experts0
On the Adversarial Robustness of Mixture of Experts0
On the effectiveness of discrete representations in sparse mixture of experts0
On the Expressive Power of Mixture-of-Experts for Structured Complex Tasks0
On the Functional Equivalence of TSK Fuzzy Systems to Neural Networks, Mixture of Experts, CART, and Stacking Ensemble Regression0
On the Representation Collapse of Sparse Mixture of Experts0
On the Risk of Evidence Pollution for Malicious Social Text Detection in the Era of LLMs0
Opportunistic Osteoporosis Diagnosis via Texture-Preserving Self-Supervision, Mixture of Experts and Multi-Task Integration0
Optimal Scaling Laws for Efficiency Gains in a Theoretical Transformer-Augmented Sectional MoE Framework0
Optimizing 6G Integrated Sensing and Communications (ISAC) via Expert Networks0
Optimizing Distributed Deployment of Mixture-of-Experts Model Inference in Serverless Computing0
Optimizing LLMs for Resource-Constrained Environments: A Survey of Model Compression Techniques0
Optimizing Mixture-of-Experts Inference Time Combining Model Deployment and Communication Scheduling0
Optimizing Mixture of Experts using Dynamic Recompilations0
Optimizing Robustness and Accuracy in Mixture of Experts: A Dual-Model Approach0
Optimus-3: Towards Generalist Multimodal Minecraft Agents with Scalable Task Experts0
P-Adapters: Robustly Extracting Factual Information from Language Models with Diverse Prompts0
Pangu Ultra MoE: How to Train Your Big MoE on Ascend NPUs0
Parameter-Efficient Conformers via Sharing Sparsely-Gated Experts for End-to-End Speech Recognition0
Parameter-Efficient Quantized Mixture-of-Experts Meets Vision-Language Instruction Tuning for Semiconductor Electron Micrograph Analysis0
Parameters vs FLOPs: Scaling Laws for Optimal Sparsity for Mixture-of-Experts Language Models0
Parameter-varying neural ordinary differential equations with partition-of-unity networks0
MoC-System: Efficient Fault Tolerance for Sparse Mixture-of-Experts Model Training0
Part-Of-Speech Sensitivity of Routers in Mixture of Experts Models0
PEMT: Multi-Task Correlation Guided Mixture-of-Experts Enables Parameter-Efficient Transfer Learning0
Perception-Informed Neural Networks: Beyond Physics-Informed Neural Networks0
PERFT: Parameter-Efficient Routed Fine-Tuning for Mixture-of-Expert Model0
Personalised Federated Learning: A Combinational Approach0
Pheromone-based Learning of Optimal Reasoning Paths0
Physics-Guided Problem Decomposition for Scaling Deep Learning of High-dimensional Eigen-Solvers: The Case of Schrödinger's Equation0
PICO: Secure Transformers via Robust Prompt Isolation and Cybersecurity Oversight0
Pipeline MoE: A Flexible MoE Implementation with Pipeline Parallelism0
Plasticity-Aware Mixture of Experts for Learning Under QoE Shifts in Adaptive Video Streaming0
PLUME: Polyhedral Learning Using Mixture of Experts0
Pluralistic Image Completion with Probabilistic Mixture-of-Experts0
Pluralistic Salient Object Detection0
P-Mamba: Marrying Perona Malik Diffusion with Mamba for Efficient Pediatric Echocardiographic Left Ventricular Segmentation0
PMoL: Parameter Efficient MoE for Preference Mixing of LLM Alignment0
Point-MoE: Towards Cross-Domain Generalization in 3D Semantic Segmentation via Mixture-of-Experts0
Polynomial-Spline Neural Networks with Exact Integrals0
Show:102550
← PrevPage 22 of 27Next →

No leaderboard results yet.