SOTAVerified

Transfer Learning

Transfer Learning is a machine learning technique where a model trained on one task is re-purposed and fine-tuned for a related, but different task. The idea behind transfer learning is to leverage the knowledge learned from a pre-trained model to solve a new, but related problem. This can be useful in situations where there is limited data available to train a new model from scratch, or when the new task is similar enough to the original task that the pre-trained model can be adapted to the new problem with only minor modifications.

( Image credit: Subodh Malgonde )

Papers

Showing 47264750 of 10307 papers

TitleStatusHype
On the impact of measure pre-conditionings on general parametric ML models and transfer learning via domain adaptation0
On the Intrinsic Limits to Representationally-Adaptive Machine-Learning0
On the Limits of Learning Representations with Label-Based Supervision0
On the Limits to Multi-Modal Popularity Prediction on Instagram -- A New Robust, Efficient and Explainable Baseline0
On the low-shot transferability of [V]-Mamba0
On the Mechanisms of Adversarial Data Augmentation for Robust and Adaptive Transfer Learning0
On the Pitfalls of Learning with Limited Data: A Facial Expression Recognition Case Study0
On The Relationship between Visual Anomaly-free and Anomalous Representations0
Unveiling the Tapestry: the Interplay of Generalization and Forgetting in Continual Learning0
On the Robustness of Arabic Speech Dialect Identification0
On the Role of Neural Collapse in Transfer Learning0
On the Role of Parallel Data in Cross-lingual Transfer Learning0
On the safety of vulnerable road users by cyclist orientation detection using Deep Learning0
On the Steganographic Capacity of Selected Learning Models0
On the Theory of Transfer Learning: The Importance of Task Diversity0
On the topology and geometry of population-based SHM0
On The Transferability of Deep-Q Networks0
On the Transferability of Massively Multilingual Pretrained Models in the Pretext of the Indo-Aryan and Tibeto-Burman Languages0
On the Transferability of Representations in Neural Networks Between Datasets and Tasks0
On the Transferability of VAE Embeddings using Relational Knowledge with Semi-Supervision0
On the Transfer of Knowledge in Quantum Algorithms0
On the universality of neural encodings in CNNs0
Analysis of Knowledge Transfer in Kernel Regime0
On the Usability of Transformers-based models for a French Question-Answering task0
On the Use of Power Amplifier Nonlinearity Quotient to Improve Radio Frequency Fingerprint Identification in Time-Varying Channels0
Show:102550
← PrevPage 190 of 413Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1APCLIPAccuracy84.2Unverified
2DFA-ENTAccuracy69.2Unverified
3DFA-SAFNAccuracy69.1Unverified
4EasyTLAccuracy63.3Unverified
5MEDAAccuracy60.3Unverified
#ModelMetricClaimedVerifiedStatus
1CNN10-20% Mask PSNR3.23Unverified
#ModelMetricClaimedVerifiedStatus
1Chatterjee, Dutta et al.[1]Accuracy96.12Unverified
#ModelMetricClaimedVerifiedStatus
1Co-TuningAccuracy85.65Unverified
#ModelMetricClaimedVerifiedStatus
1Physical AccessEER5.74Unverified
#ModelMetricClaimedVerifiedStatus
1riadd.aucmediAUROC0.95Unverified