SOTAVerified

Transfer Learning

Transfer Learning is a machine learning technique where a model trained on one task is re-purposed and fine-tuned for a related, but different task. The idea behind transfer learning is to leverage the knowledge learned from a pre-trained model to solve a new, but related problem. This can be useful in situations where there is limited data available to train a new model from scratch, or when the new task is similar enough to the original task that the pre-trained model can be adapted to the new problem with only minor modifications.

( Image credit: Subodh Malgonde )

Papers

Showing 16261650 of 10307 papers

TitleStatusHype
DeviceBERT: Applied Transfer Learning With Targeted Annotations and Vocabulary Enrichment to Identify Medical Device and Component Terminology in FDA Recall Summaries0
Online Policy Distillation with Decision-Attention0
Accelerating evolutionary exploration through language model-based transfer learning0
Low-Resource Cross-Lingual Summarization through Few-Shot Learning with Large Language Models0
FunBO: Discovering Acquisition Functions for Bayesian Optimization with FunSearch0
Through the Thicket: A Study of Number-Oriented LLMs derived from Random Forest Models0
Labeled Data Selection for Category Discovery0
InaGVAD : a Challenging French TV and Radio Corpus Annotated for Speech Activity Detection and Speaker Gender SegmentationCode1
Continuous Test-time Domain Adaptation for Efficient Fault Detection under Evolving Operating Conditions0
LLMEmbed: Rethinking Lightweight LLM's Genuine Function in Text ClassificationCode1
Step Out and Seek Around: On Warm-Start Training with Incremental Data0
UrbanSARFloods: Sentinel-1 SLC-Based Benchmark Dataset for Urban and Open-Area Flood Mapping0
Attend and Enrich: Enhanced Visual Prompt for Zero-Shot Learning0
Exploring User Retrieval Integration towards Large Language Models for Cross-Domain Sequential RecommendationCode0
Multi-Task Multi-Scale Contrastive Knowledge Distillation for Efficient Medical Image SegmentationCode1
Transfer Learning for Latent Variable Network Models0
M2D-CLAP: Masked Modeling Duo Meets CLAP for Learning General-purpose Audio-Language Representation0
Randomized Geometric Algebra Methods for Convex Neural NetworksCode0
Leveraging Predicate and Triplet Learning for Scene Graph GenerationCode1
Towards Neural Architecture Search for Transfer Learning in 6G Networks0
CADE: Cosine Annealing Differential Evolution for Spiking Neural NetworkCode0
Enhancing Trust in LLMs: Algorithms for Comparing and Interpreting LLMs0
FedMKT: Federated Mutual Knowledge Transfer for Large and Small Language ModelsCode3
Multi-Agent Transfer Learning via Temporal Contrastive Learning0
The Empirical Impact of Forgetting and Transfer in Continual Visual Odometry0
Show:102550
← PrevPage 66 of 413Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1APCLIPAccuracy84.2Unverified
2DFA-ENTAccuracy69.2Unverified
3DFA-SAFNAccuracy69.1Unverified
4EasyTLAccuracy63.3Unverified
5MEDAAccuracy60.3Unverified
#ModelMetricClaimedVerifiedStatus
1CNN10-20% Mask PSNR3.23Unverified
#ModelMetricClaimedVerifiedStatus
1Chatterjee, Dutta et al.[1]Accuracy96.12Unverified
#ModelMetricClaimedVerifiedStatus
1Co-TuningAccuracy85.65Unverified
#ModelMetricClaimedVerifiedStatus
1Physical AccessEER5.74Unverified
#ModelMetricClaimedVerifiedStatus
1riadd.aucmediAUROC0.95Unverified