SOTAVerified

Contrastive Learning

Contrastive Learning is a deep learning technique for unsupervised representation learning. The goal is to learn a representation of data such that similar instances are close together in the representation space, while dissimilar instances are far apart.

It has been shown to be effective in various computer vision and natural language processing tasks, including image retrieval, zero-shot learning, and cross-modal retrieval. In these tasks, the learned representations can be used as features for downstream tasks such as classification and clustering.

(Image credit: Schroff et al. 2015)

Papers

Showing 226250 of 6661 papers

TitleStatusHype
CoCo: Coherence-Enhanced Machine-Generated Text Detection Under Data Limitation With Contrastive LearningCode1
CoCoNet: Coupled Contrastive Learning Network with Multi-level Feature Ensemble for Multi-modality Image FusionCode1
CoMAE: Single Model Hybrid Pre-training on Small-Scale RGB-D DatasetsCode1
Composed Image Retrieval using Contrastive Learning and Task-oriented CLIP-based FeaturesCode1
Anatomical Invariance Modeling and Semantic Alignment for Self-supervised Learning in 3D Medical Image AnalysisCode1
A Language Model based Framework for New Concept Placement in OntologiesCode1
CODE: Contrastive Pre-training with Adversarial Fine-tuning for Zero-shot Expert LinkingCode1
COARSE3D: Class-Prototypes for Contrastive Learning in Weakly-Supervised 3D Point Cloud SegmentationCode1
Co^2L: Contrastive Continual LearningCode1
ACTION++: Improving Semi-supervised Medical Image Segmentation with Adaptive Anatomical ContrastCode1
Co2L: Contrastive Continual LearningCode1
A latent space for unsupervised MR image quality control via artifact assessmentCode1
Actionness Inconsistency-guided Contrastive Learning for Weakly-supervised Temporal Action LocalizationCode1
AIRCHITECT v2: Learning the Hardware Accelerator Design Space through Unified RepresentationsCode1
An Empirical Study on Disentanglement of Negative-free Contrastive LearningCode1
Anatomical Foundation Models for Brain MRIsCode1
COCO-LM: Correcting and Contrasting Text Sequences for Language Model PretrainingCode1
CoCoNets: Continuous Contrastive 3D Scene RepresentationsCode1
CODER: Knowledge infused cross-lingual medical term embedding for term normalizationCode1
CoIn: Contrastive Instance Feature Mining for Outdoor 3D Object Detection with Very Limited AnnotationsCode1
CoLA: Weakly-Supervised Temporal Action Localization with Snippet Contrastive LearningCode1
Aligning Language Models with Human Preferences via a Bayesian ApproachCode1
Breaking the Batch Barrier (B3) of Contrastive Learning via Smart Batch MiningCode1
Aligning Pretraining for Detection via Object-Level Contrastive LearningCode1
CMID: A Unified Self-Supervised Learning Framework for Remote Sensing Image UnderstandingCode1
Show:102550
← PrevPage 10 of 267Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50ImageNet Top-1 Accuracy73.6Unverified
2ResNet50ImageNet Top-1 Accuracy73Unverified
3ResNet50ImageNet Top-1 Accuracy71.1Unverified
4ResNet50ImageNet Top-1 Accuracy69.3Unverified
5ResNet50 (v2)ImageNet Top-1 Accuracy67.6Unverified
6ResNet50 (v2)ImageNet Top-1 Accuracy63.8Unverified
7ResNet50ImageNet Top-1 Accuracy63.6Unverified
8ResNet50ImageNet Top-1 Accuracy61.5Unverified
9ResNet50ImageNet Top-1 Accuracy61.5Unverified
10ResNet50 (4×)ImageNet Top-1 Accuracy61.3Unverified
#ModelMetricClaimedVerifiedStatus
110..5sec1Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)84.77Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)85.55Unverified