SOTAVerified

Contrastive Learning

Contrastive Learning is a deep learning technique for unsupervised representation learning. The goal is to learn a representation of data such that similar instances are close together in the representation space, while dissimilar instances are far apart.

It has been shown to be effective in various computer vision and natural language processing tasks, including image retrieval, zero-shot learning, and cross-modal retrieval. In these tasks, the learned representations can be used as features for downstream tasks such as classification and clustering.

(Image credit: Schroff et al. 2015)

Papers

Showing 48264850 of 6661 papers

TitleStatusHype
Evaluation of Contrastive Learning with Various Code Representations for Code Clone Detection0
MET: Masked Encoding for Tabular DataCode1
DU-Net based Unsupervised Contrastive Learning for Cancer Segmentation in Histology Images0
Learning Fair Representation via Distributional Contrastive DisentanglementCode1
VLMixer: Unpaired Vision-Language Pre-training via Cross-Modal CutMix0
Beyond Supervised vs. Unsupervised: Representative Benchmarking and Analysis of Image Representation LearningCode0
NCAGC: A Neighborhood Contrast Framework for Attributed Graph ClusteringCode1
Volumetric Supervised Contrastive Learning for Seismic Semantic Segmentation0
Let Invariant Rationale Discovery Inspire Graph Contrastive LearningCode1
Text-Aware End-to-end Mispronunciation Detection and DiagnosisCode1
Contrastive Learning as Goal-Conditioned Reinforcement Learning0
Discrete Contrastive Diffusion for Cross-Modal Music and Image GenerationCode1
Label-enhanced Prototypical Network with Contrastive Learning for Multi-label Few-shot Aspect Category Detection0
Towards a Solution to Bongard Problems: A Causal Approach0
On Finite-Sample Identifiability of Contrastive Learning-Based Nonlinear Independent Component Analysis0
Self-Supervised Representation Learning With MUlti-Segmental Informational Coding (MUSIC)0
Contrastive Learning for Unsupervised Domain Adaptation of Time SeriesCode1
Transductive CLIP with Class-Conditional Contrastive Learning0
GLIPv2: Unifying Localization and Vision-Language UnderstandingCode4
Is Self-Supervised Learning More Robust Than Supervised Learning?0
ClamNet: Using contrastive learning with variable depth Unets for medical image segmentation0
STNDT: Modeling Neural Population Activity with a Spatiotemporal Transformer0
COSTA: Covariance-Preserving Feature Augmentation for Graph Contrastive LearningCode1
I'm Me, We're Us, and I'm Us: Tri-directional Contrastive Learning on HypergraphsCode1
An Empirical Study on Disentanglement of Negative-free Contrastive LearningCode1
Show:102550
← PrevPage 194 of 267Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50ImageNet Top-1 Accuracy73.6Unverified
2ResNet50ImageNet Top-1 Accuracy73Unverified
3ResNet50ImageNet Top-1 Accuracy71.1Unverified
4ResNet50ImageNet Top-1 Accuracy69.3Unverified
5ResNet50 (v2)ImageNet Top-1 Accuracy67.6Unverified
6ResNet50 (v2)ImageNet Top-1 Accuracy63.8Unverified
7ResNet50ImageNet Top-1 Accuracy63.6Unverified
8ResNet50ImageNet Top-1 Accuracy61.5Unverified
9ResNet50ImageNet Top-1 Accuracy61.5Unverified
10ResNet50 (4×)ImageNet Top-1 Accuracy61.3Unverified
#ModelMetricClaimedVerifiedStatus
110..5sec1Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)84.77Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)85.55Unverified