SOTAVerified

Multi-Task Learning

Multi-task learning aims to learn multiple different tasks simultaneously while maximizing performance on one or all of the tasks.

( Image credit: Cross-stitch Networks for Multi-task Learning )

Papers

Showing 125 of 3687 papers

TitleStatusHype
A Comprehensive Survey of Mixture-of-Experts: Algorithms, Theory, and ApplicationsCode9
Arcee's MergeKit: A Toolkit for Merging Large Language ModelsCode9
VITA: Towards Open-Source Interactive Omni Multimodal LLMCode7
Gradient-Based Multi-Objective Deep Learning: Algorithms, Theories, Applications, and BeyondCode7
MiniGPT-v2: large language model as a unified interface for vision-language multi-task learningCode7
StreamSpeech: Simultaneous Speech-to-Speech Translation with Multi-task LearningCode5
MING-MOE: Enhancing Medical Multi-Task Learning in Large Language Models with Sparse Mixture of Low-Rank Adapter ExpertsCode5
YOLOR-Based Multi-Task LearningCode5
Uni-Perceiver v2: A Generalist Model for Large-Scale Vision and Vision-Language TasksCode4
CoBa: Convergence Balancer for Multitask Finetuning of Large Language ModelsCode4
Model Merging in LLMs, MLLMs, and Beyond: Methods, Theories, Applications and OpportunitiesCode4
InternLM-Math: Open Math Large Language Models Toward Verifiable ReasoningCode4
DocRes: A Generalist Model Toward Unifying Document Image Restoration TasksCode4
Gradient Alignment in Physics-informed Neural Networks: A Second-Order Optimization PerspectiveCode3
Scientific Machine Learning through Physics-Informed Neural Networks: Where we are and What's nextCode3
UCF: Uncovering Common Features for Generalizable Deepfake DetectionCode3
MVMoE: Multi-Task Vehicle Routing Solver with Mixture-of-ExpertsCode3
DARWIN 1.5: Large Language Models as Materials Science Adapted LearnersCode3
ERNIE 2.0: A Continual Pre-training Framework for Language UnderstandingCode3
PETRv2: A Unified Framework for 3D Perception from Multi-Camera ImagesCode3
Ludwig: a type-based declarative deep learning toolboxCode3
Language Models are Few-Shot LearnersCode3
MixLoRA: Enhancing Large Language Models Fine-Tuning with LoRA-based Mixture of ExpertsCode3
Relational Multi-Task Learning: Modeling Relations between Data and TasksCode3
YourMT3+: Multi-instrument Music Transcription with Enhanced Transformer Architectures and Cross-dataset Stem AugmentationCode3
Show:102550
← PrevPage 1 of 148Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1PCGrad∆m%125.7Unverified
2CAGrad∆m%112.8Unverified
3IMTL-G∆m%77.2Unverified
4Nash-MTL∆m%62Unverified
5BayesAgg-MTL∆m%53.7Unverified
#ModelMetricClaimedVerifiedStatus
1SwinMTLmIoU76.41Unverified
2Nash-MTLmIoU75.41Unverified
3MultiObjectiveOptimizationmIoU66.63Unverified
#ModelMetricClaimedVerifiedStatus
1SwinMTLMean IoU58.14Unverified
2Nash-MTLMean IoU40.13Unverified
#ModelMetricClaimedVerifiedStatus
1Gumbel-Matrix RoutingAverage Accuracy93.52Unverified
2Mixture-of-ExpertsAverage Accuracy92.19Unverified
#ModelMetricClaimedVerifiedStatus
1MGDA-UBError8.25Unverified
#ModelMetricClaimedVerifiedStatus
1BayesAgg-MTLdelta_m-2.23Unverified
#ModelMetricClaimedVerifiedStatus
1LETRFH83.3Unverified