SOTAVerified

Contrastive Learning

Contrastive Learning is a deep learning technique for unsupervised representation learning. The goal is to learn a representation of data such that similar instances are close together in the representation space, while dissimilar instances are far apart.

It has been shown to be effective in various computer vision and natural language processing tasks, including image retrieval, zero-shot learning, and cross-modal retrieval. In these tasks, the learned representations can be used as features for downstream tasks such as classification and clustering.

(Image credit: Schroff et al. 2015)

Papers

Showing 18761900 of 6661 papers

TitleStatusHype
Diffusion Model is Secretly a Training-free Open Vocabulary Semantic SegmenterCode1
CluCDD:Contrastive Dialogue Disentanglement via ClusteringCode1
MetaMask: Revisiting Dimensional Confounder for Self-Supervised LearningCode1
Digging into contrastive learning for robust depth estimation with diffusion modelsCode1
CLUDA : Contrastive Learning in Unsupervised Domain Adaptation for Semantic SegmentationCode1
miCSE: Mutual Information Contrastive Learning for Low-shot Sentence EmbeddingsCode1
MET: Masked Encoding for Tabular DataCode1
MHCCL: Masked Hierarchical Cluster-Wise Contrastive Learning for Multivariate Time SeriesCode1
Extending global-local view alignment for self-supervised learning with remote sensing imageryCode1
Directed Graph Contrastive LearningCode1
MiCE: Mixture of Contrastive Experts for Unsupervised Image ClusteringCode1
AdaGCL: Adaptive Subgraph Contrastive Learning to Generalize Large-scale Graph TrainingCode0
Alignist: CAD-Informed Orientation Distribution Estimation by Fusing Shape and CorrespondencesCode0
AdaCrossNet: Adaptive Dynamic Loss Weighting for Cross-Modal Contrastive Point Cloud LearningCode0
Multichannel AV-wav2vec2: A Framework for Learning Multichannel Multi-Modal Speech RepresentationCode0
Multi-Graph Co-Training for Capturing User Intent in Session-based RecommendationCode0
MuDAF: Long-Context Multi-Document Attention Focusing through Contrastive Learning on Attention HeadsCode0
A Unified Contrastive Loss for Self-TrainingCode0
Aligning Visual Contrastive learning models via Preference OptimizationCode0
MTS-LOF: Medical Time-Series Representation Learning via Occlusion-Invariant FeaturesCode0
Collaborative Unsupervised Visual Representation Learning from Decentralized DataCode0
CGCL: Collaborative Graph Contrastive Learning without Handcrafted Graph Data AugmentationsCode0
A Unified and Scalable Membership Inference Method for Visual Self-supervised Encoder via Part-aware CapabilityCode0
MSCDA: Multi-level Semantic-guided Contrast Improves Unsupervised Domain Adaptation for Breast MRI Segmentation in Small DatasetsCode0
MSVQ: Self-Supervised Learning with Multiple Sample Views and QueuesCode0
Show:102550
← PrevPage 76 of 267Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50ImageNet Top-1 Accuracy73.6Unverified
2ResNet50ImageNet Top-1 Accuracy73Unverified
3ResNet50ImageNet Top-1 Accuracy71.1Unverified
4ResNet50ImageNet Top-1 Accuracy69.3Unverified
5ResNet50 (v2)ImageNet Top-1 Accuracy67.6Unverified
6ResNet50 (v2)ImageNet Top-1 Accuracy63.8Unverified
7ResNet50ImageNet Top-1 Accuracy63.6Unverified
8ResNet50ImageNet Top-1 Accuracy61.5Unverified
9ResNet50ImageNet Top-1 Accuracy61.5Unverified
10ResNet50 (4×)ImageNet Top-1 Accuracy61.3Unverified
#ModelMetricClaimedVerifiedStatus
110..5sec1Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)84.77Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)85.55Unverified