SOTAVerified

Transfer Learning

Transfer Learning is a machine learning technique where a model trained on one task is re-purposed and fine-tuned for a related, but different task. The idea behind transfer learning is to leverage the knowledge learned from a pre-trained model to solve a new, but related problem. This can be useful in situations where there is limited data available to train a new model from scratch, or when the new task is similar enough to the original task that the pre-trained model can be adapted to the new problem with only minor modifications.

( Image credit: Subodh Malgonde )

Papers

Showing 16261650 of 10307 papers

TitleStatusHype
A Review and Implementation of Object Detection Models and Optimizations for Real-time Medical Mask Detection during the COVID-19 PandemicCode0
Are Structural Concepts Universal in Transformer Language Models? Towards Interpretable Cross-Lingual GeneralizationCode0
A Resource-Efficient Training Framework for Remote Sensing Text--Image RetrievalCode0
Human Genome Book: Words, Sentences and ParagraphsCode0
Human-Inspired Framework to Accelerate Reinforcement LearningCode0
Hyperparameters in Score-Based Membership Inference AttacksCode0
A Regularization-based Transfer Learning Method for Information Extraction via Instructed Graph DecoderCode0
HTR-JAND: Handwritten Text Recognition with Joint Attention Network and Knowledge DistillationCode0
Are ECGs enough? Deep learning classification of cardiac anomalies using only electrocardiogramsCode0
HR-VILAGE-3K3M: A Human Respiratory Viral Immunization Longitudinal Gene Expression Dataset for Systems ImmunityCode0
hULMonA: The Universal Language Model in ArabicCode0
How Well Do Vision Transformers (VTs) Transfer To The Non-Natural Image Domain? An Empirical Study Involving Art ClassificationCode0
Hyperpolyglot LLMs: Cross-Lingual Interpretability in Token EmbeddingsCode0
Implicit Cross-Lingual Rewarding for Efficient Multilingual Preference AlignmentCode0
How should we evaluate supervised hashing?Code0
How to evaluate word embeddings? On importance of data efficiency and simple supervised tasksCode0
Arabic Dialect Identification Using BERT Fine-TuningCode0
How Language-Neutral is Multilingual BERT?Code0
How to tackle an emerging topic? Combining strong and weak labels for Covid news NERCode0
Training-Free Acceleration of ViTs with Delayed Spatial MergingCode0
How good are variational autoencoders at transfer learning?Code0
Accelerating Transfer Learning with Near-Data Computation on Cloud Object StoresCode0
How does Multi-Task Training Affect Transformer In-Context Capabilities? Investigations with Function ClassesCode0
How to Train a CAT: Learning Canonical Appearance Transformations for Direct Visual Localization Under Illumination ChangeCode0
Aesthetic Attributes Assessment of ImagesCode0
Show:102550
← PrevPage 66 of 413Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1APCLIPAccuracy84.2Unverified
2DFA-ENTAccuracy69.2Unverified
3DFA-SAFNAccuracy69.1Unverified
4EasyTLAccuracy63.3Unverified
5MEDAAccuracy60.3Unverified
#ModelMetricClaimedVerifiedStatus
1CNN10-20% Mask PSNR3.23Unverified
#ModelMetricClaimedVerifiedStatus
1Chatterjee, Dutta et al.[1]Accuracy96.12Unverified
#ModelMetricClaimedVerifiedStatus
1Co-TuningAccuracy85.65Unverified
#ModelMetricClaimedVerifiedStatus
1Physical AccessEER5.74Unverified
#ModelMetricClaimedVerifiedStatus
1riadd.aucmediAUROC0.95Unverified