SOTAVerified

Contrastive Learning

Contrastive Learning is a deep learning technique for unsupervised representation learning. The goal is to learn a representation of data such that similar instances are close together in the representation space, while dissimilar instances are far apart.

It has been shown to be effective in various computer vision and natural language processing tasks, including image retrieval, zero-shot learning, and cross-modal retrieval. In these tasks, the learned representations can be used as features for downstream tasks such as classification and clustering.

(Image credit: Schroff et al. 2015)

Papers

Showing 32513275 of 6661 papers

TitleStatusHype
Zero-shot domain adaptation based on dual-level mix and contrast0
Leveraging Contrastive Learning for Enhanced Node Representations in Tokenized Graph Transformers0
ProtoGMM: Multi-prototype Gaussian-Mixture-based Domain Adaptation Model for Semantic Segmentation0
Improving the Consistency in Cross-Lingual Cross-Modal Retrieval with 1-to-K Contrastive LearningCode0
Foundation Models for ECG: Leveraging Hybrid Self-Supervised Learning for Advanced Cardiac Diagnostics0
Investigating Self-Supervised Methods for Label-Efficient Learning0
TopoGCL: Topological Graph Contrastive LearningCode0
Data curation via joint example selection further accelerates multimodal learning0
Hyperbolic Knowledge Transfer in Cross-Domain Recommendation System0
Contrastive General Graph Matching with Adaptive Augmentation Sampling0
MPCODER: Multi-user Personalized Code Generator with Explicit and Implicit Style Representation LearningCode0
Enhancing OOD Detection Using Latent DiffusionCode0
The Championship-Winning Solution for the 5th CLVISION Challenge 20240
Exploring Test-Time Adaptation for Object Detection in Continually Changing Environments0
Self-Supervised Alignment Learning for Medical Image Segmentation0
Fine-grained Background Representation for Weakly Supervised Semantic SegmentationCode0
Speech Analysis of Language Varieties in ItalyCode0
DN-CL: Deep Symbolic Regression against Noise via Contrastive Learning0
TemPrompt: Multi-Task Prompt Learning for Temporal Relation Extraction in RAG-based Crowdsourcing Systems0
From Overfitting to Robustness: Quantity, Quality, and Variety Oriented Negative Sample Selection in Graph Contrastive Learning0
Enhancing Idiomatic Representation in Multiple Languages via an Adaptive Contrastive Triplet Loss0
A Contrastive Learning Approach to Mitigate Bias in Speech ModelsCode0
Revealing Vision-Language Integration in the Brain with Multimodal NetworksCode0
Maintenance Required: Updating and Extending Bootstrapped Human Activity Recognition Systems for Smart Homes0
Factual Dialogue Summarization via Learning from Large Language Models0
Show:102550
← PrevPage 131 of 267Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50ImageNet Top-1 Accuracy73.6Unverified
2ResNet50ImageNet Top-1 Accuracy73Unverified
3ResNet50ImageNet Top-1 Accuracy71.1Unverified
4ResNet50ImageNet Top-1 Accuracy69.3Unverified
5ResNet50 (v2)ImageNet Top-1 Accuracy67.6Unverified
6ResNet50 (v2)ImageNet Top-1 Accuracy63.8Unverified
7ResNet50ImageNet Top-1 Accuracy63.6Unverified
8ResNet50ImageNet Top-1 Accuracy61.5Unverified
9ResNet50ImageNet Top-1 Accuracy61.5Unverified
10ResNet50 (4×)ImageNet Top-1 Accuracy61.3Unverified
#ModelMetricClaimedVerifiedStatus
110..5sec1Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)84.77Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)85.55Unverified