SOTAVerified

Contrastive Learning

Contrastive Learning is a deep learning technique for unsupervised representation learning. The goal is to learn a representation of data such that similar instances are close together in the representation space, while dissimilar instances are far apart.

It has been shown to be effective in various computer vision and natural language processing tasks, including image retrieval, zero-shot learning, and cross-modal retrieval. In these tasks, the learned representations can be used as features for downstream tasks such as classification and clustering.

(Image credit: Schroff et al. 2015)

Papers

Showing 13511375 of 6661 papers

TitleStatusHype
Improving Word Translation via Two-Stage Contrastive LearningCode1
Improving Event Representation via Simultaneous Weakly Supervised Contrastive Learning and ClusteringCode1
Improved Universal Sentence Embeddings with Prompt-based Contrastive Learning and Energy-based LearningCode1
Rethinking Minimal Sufficient Representation in Contrastive LearningCode1
S5CL: Unifying Fully-Supervised, Self-Supervised, and Semi-Supervised Learning Through Hierarchical Contrastive LearningCode1
LaPraDoR: Unsupervised Pretrained Dense Retriever for Zero-Shot Text RetrievalCode1
Neuromorphic Data Augmentation for Training Spiking Neural NetworksCode1
Democracy Does Matter: Comprehensive Feature Mining for Co-Salient Object DetectionCode1
A Sentence is Worth 128 Pseudo Tokens: A Semantic-Aware Contrastive Learning Framework for Sentence EmbeddingsCode1
MetAug: Contrastive Learning via Meta Feature AugmentationCode1
Domain Generalization via Shuffled Style Assembly for Face Anti-SpoofingCode1
What Matters For Meta-Learning Vision Regression Tasks?Code1
Incorporating Hierarchy into Text Encoder: a Contrastive Learning Approach for Hierarchical Text ClassificationCode1
Selective-Supervised Contrastive Learning with Noisy LabelsCode1
UniXcoder: Unified Cross-Modal Pre-training for Code RepresentationCode1
Interpretable part-whole hierarchies and conceptual-semantic relationships in neural networksCode1
Consistent Representation Learning for Continual Relation ExtractionCode1
Exploring Patch-wise Semantic Relation for Contrastive Learning in Image-to-Image Translation TasksCode1
Mind the Gap: Understanding the Modality Gap in Multi-modal Contrastive Representation LearningCode1
Correct-N-Contrast: A Contrastive Approach for Improving Robustness to Spurious CorrelationsCode1
MERIt: Meta-Path Guided Contrastive Learning for Logical ReasoningCode1
Domain Knowledge-Informed Self-Supervised Representations for Workout Form AssessmentCode1
UCTopic: Unsupervised Contrastive Learning for Phrase Representations and Topic MiningCode1
HiCLRE: A Hierarchical Contrastive Learning Framework for Distantly Supervised Relation ExtractionCode1
Disentangling Long and Short-Term Interests for RecommendationCode1
Show:102550
← PrevPage 55 of 267Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50ImageNet Top-1 Accuracy73.6Unverified
2ResNet50ImageNet Top-1 Accuracy73Unverified
3ResNet50ImageNet Top-1 Accuracy71.1Unverified
4ResNet50ImageNet Top-1 Accuracy69.3Unverified
5ResNet50 (v2)ImageNet Top-1 Accuracy67.6Unverified
6ResNet50 (v2)ImageNet Top-1 Accuracy63.8Unverified
7ResNet50ImageNet Top-1 Accuracy63.6Unverified
8ResNet50ImageNet Top-1 Accuracy61.5Unverified
9ResNet50ImageNet Top-1 Accuracy61.5Unverified
10ResNet50 (4×)ImageNet Top-1 Accuracy61.3Unverified
#ModelMetricClaimedVerifiedStatus
110..5sec1Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)84.77Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)85.55Unverified