SOTAVerified

Transfer Learning

Transfer Learning is a machine learning technique where a model trained on one task is re-purposed and fine-tuned for a related, but different task. The idea behind transfer learning is to leverage the knowledge learned from a pre-trained model to solve a new, but related problem. This can be useful in situations where there is limited data available to train a new model from scratch, or when the new task is similar enough to the original task that the pre-trained model can be adapted to the new problem with only minor modifications.

( Image credit: Subodh Malgonde )

Papers

Showing 526550 of 10307 papers

TitleStatusHype
Are You Stealing My Model? Sample Correlation for Fingerprinting Deep Neural NetworksCode1
Pre-training technique to localize medical BERT and enhance biomedical BERTCode1
CovidCTNet: An Open-Source Deep Learning Approach to Identify Covid-19 Using CT ImageCode1
A proposal for Multimodal Emotion Recognition using aural transformers and Action Units on RAVDESS datasetCode1
A Comparative Study of Existing and New Deep Learning Methods for Detecting Knee Injuries using the MRNet DatasetCode1
AquilaMoE: Efficient Training for MoE Models with Scale-Up and Scale-Out StrategiesCode1
COVID-CXNet: Detecting COVID-19 in Frontal Chest X-ray Images using Deep LearningCode1
CrAM: A Compression-Aware MinimizerCode1
Convolutional Bypasses Are Better Vision Transformer AdaptersCode1
A Comparative Study of Deep Reinforcement Learning-based Transferable Energy Management Strategies for Hybrid Electric VehiclesCode1
Convolutional Neural Networks for Classification of Alzheimer's Disease: Overview and Reproducible EvaluationCode1
ConvLab-3: A Flexible Dialogue System Toolkit Based on a Unified Data FormatCode1
Conv-Adapter: Exploring Parameter Efficient Transfer Learning for ConvNetsCode1
ConvNet vs Transformer, Supervised vs CLIP: Beyond ImageNet AccuracyCode1
Cooperative Self-training of Machine Reading ComprehensionCode1
AP-10K: A Benchmark for Animal Pose Estimation in the WildCode1
3D Point Cloud Registration with Multi-Scale Architecture and Unsupervised Transfer LearningCode1
AnyStar: Domain randomized universal star-convex 3D instance segmentationCode1
Contrastive Representation DistillationCode1
Co-Tuning for Transfer LearningCode1
An Uncertainty-aware Transfer Learning-based Framework for Covid-19 DiagnosisCode1
Contrastive Alignment of Vision to Language Through Parameter-Efficient Transfer LearningCode1
Contrastive Cross-domain Recommendation in MatchingCode1
Continual Sequence Generation with Adaptive Compositional ModulesCode1
Contour Knowledge Transfer for Salient Object DetectionCode1
Show:102550
← PrevPage 22 of 413Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1APCLIPAccuracy84.2Unverified
2DFA-ENTAccuracy69.2Unverified
3DFA-SAFNAccuracy69.1Unverified
4EasyTLAccuracy63.3Unverified
5MEDAAccuracy60.3Unverified
#ModelMetricClaimedVerifiedStatus
1CNN10-20% Mask PSNR3.23Unverified
#ModelMetricClaimedVerifiedStatus
1Chatterjee, Dutta et al.[1]Accuracy96.12Unverified
#ModelMetricClaimedVerifiedStatus
1Co-TuningAccuracy85.65Unverified
#ModelMetricClaimedVerifiedStatus
1Physical AccessEER5.74Unverified
#ModelMetricClaimedVerifiedStatus
1riadd.aucmediAUROC0.95Unverified