SOTAVerified

Contrastive Learning

Contrastive Learning is a deep learning technique for unsupervised representation learning. The goal is to learn a representation of data such that similar instances are close together in the representation space, while dissimilar instances are far apart.

It has been shown to be effective in various computer vision and natural language processing tasks, including image retrieval, zero-shot learning, and cross-modal retrieval. In these tasks, the learned representations can be used as features for downstream tasks such as classification and clustering.

(Image credit: Schroff et al. 2015)

Papers

Showing 19762000 of 6661 papers

TitleStatusHype
Multi-Stage Contrastive Regression for Action Quality AssessmentCode0
MuDAF: Long-Context Multi-Document Attention Focusing through Contrastive Learning on Attention HeadsCode0
Disentangled Contrastive Learning for Learning Robust Textual RepresentationsCode0
MTS-LOF: Medical Time-Series Representation Learning via Occlusion-Invariant FeaturesCode0
Multichannel AV-wav2vec2: A Framework for Learning Multichannel Multi-Modal Speech RepresentationCode0
CM3AE: A Unified RGB Frame and Event-Voxel/-Frame Pre-training FrameworkCode0
Domain Generalization for Text Classification with Memory-Based Supervised Contrastive LearningCode0
CL-XABSA: Contrastive Learning for Cross-lingual Aspect-based Sentiment AnalysisCode0
ALBRT: Cellular Composition Prediction in Routine Histology ImagesCode0
Discriminative Representation learning via Attention-Enhanced Contrastive Learning for Short Text ClusteringCode0
MSA-UNet3+: Multi-Scale Attention UNet3+ with New Supervised Prototypical Contrastive Loss for Coronary DSA Image SegmentationCode0
MSCDA: Multi-level Semantic-guided Contrast Improves Unsupervised Domain Adaptation for Breast MRI Segmentation in Small DatasetsCode0
MSVQ: Self-Supervised Learning with Multiple Sample Views and QueuesCode0
Motif-Centric Representation Learning for Symbolic MusicCode0
Discovering Global False Negatives On the Fly for Self-supervised Contrastive LearningCode0
Morality is Non-Binary: Building a Pluralist Moral Sentence Embedding Space using Contrastive LearningCode0
M(otion)-mode Based Prediction of Ejection Fraction using EchocardiogramsCode0
MoMA: Momentum Contrastive Learning with Multi-head Attention-based Knowledge Distillation for Histopathology Image AnalysisCode0
DISCO: A Hierarchical Disentangled Cognitive Diagnosis Framework for Interpretable Job RecommendationCode0
A Language-based solution to enable Metaverse RetrievalCode0
Disambiguated Node Classification with Graph Neural NetworksCode0
Cluster-guided Asymmetric Contrastive Learning for Unsupervised Person Re-IdentificationCode0
Molecular Graph Contrastive Learning with Line GraphCode0
MolPLA: A Molecular Pretraining Framework for Learning Cores, R-Groups and their Linker JointsCode0
MOOSS: Mask-Enhanced Temporal Contrastive Learning for Smooth State Evolution in Visual Reinforcement LearningCode0
Show:102550
← PrevPage 80 of 267Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50ImageNet Top-1 Accuracy73.6Unverified
2ResNet50ImageNet Top-1 Accuracy73Unverified
3ResNet50ImageNet Top-1 Accuracy71.1Unverified
4ResNet50ImageNet Top-1 Accuracy69.3Unverified
5ResNet50 (v2)ImageNet Top-1 Accuracy67.6Unverified
6ResNet50 (v2)ImageNet Top-1 Accuracy63.8Unverified
7ResNet50ImageNet Top-1 Accuracy63.6Unverified
8ResNet50ImageNet Top-1 Accuracy61.5Unverified
9ResNet50ImageNet Top-1 Accuracy61.5Unverified
10ResNet50 (4×)ImageNet Top-1 Accuracy61.3Unverified
#ModelMetricClaimedVerifiedStatus
110..5sec1Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)84.77Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)85.55Unverified