SOTAVerified

Contrastive Learning

Contrastive Learning is a deep learning technique for unsupervised representation learning. The goal is to learn a representation of data such that similar instances are close together in the representation space, while dissimilar instances are far apart.

It has been shown to be effective in various computer vision and natural language processing tasks, including image retrieval, zero-shot learning, and cross-modal retrieval. In these tasks, the learned representations can be used as features for downstream tasks such as classification and clustering.

(Image credit: Schroff et al. 2015)

Papers

Showing 5175 of 6661 papers

TitleStatusHype
Exploring Contrastive Learning for Multimodal Detection of Misogynistic MemesCode2
Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in AlignmentCode2
A Systematic Study of Joint Representation Learning on Protein Sequences and StructuresCode2
Extended Agriculture-Vision: An Extension of a Large Aerial Image Dataset for Agricultural Pattern AnalysisCode2
FontDiffuser: One-Shot Font Generation via Denoising Diffusion with Multi-Scale Content Aggregation and Style Contrastive LearningCode2
Enhancing Multi-view Stereo with Contrastive Matching and Weighted Focal LossCode2
Analyzing and Boosting the Power of Fine-Grained Visual Recognition for Multi-modal Large Language ModelsCode2
4D Contrastive Superflows are Dense 3D Representation LearnersCode2
Egocentric Video-Language PretrainingCode2
EyeCLIP: A visual-language foundation model for multi-modal ophthalmic image analysisCode2
EasyRec: Simple yet Effective Language Models for RecommendationCode2
Few-Shot Scene Classification of Optical Remote Sensing Images Leveraging Calibrated Pretext TasksCode2
DreamLIP: Language-Image Pre-training with Long CaptionsCode2
ECG-Chat: A Large ECG-Language Model for Cardiac Disease DiagnosisCode2
End-to-end Learnable Clustering for Intent Learning in RecommendationCode2
DiffCSE: Difference-based Contrastive Learning for Sentence EmbeddingsCode2
DiffMM: Multi-Modal Diffusion Model for RecommendationCode2
Detecting and Grounding Multi-Modal Media Manipulation and BeyondCode2
Detecting and Grounding Multi-Modal Media ManipulationCode2
DeTeCtive: Detecting AI-generated Text via Multi-Level Contrastive LearningCode2
DNABERT-S: Pioneering Species Differentiation with Species-Aware DNA EmbeddingsCode2
Enhanced Contrastive Learning with Multi-view Longitudinal Data for Chest X-ray Report GenerationCode2
Decoupling Static and Hierarchical Motion Perception for Referring Video SegmentationCode2
Decoding speech perception from non-invasive brain recordingsCode2
Delving into Inter-Image Invariance for Unsupervised Visual RepresentationsCode2
Show:102550
← PrevPage 3 of 267Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50ImageNet Top-1 Accuracy73.6Unverified
2ResNet50ImageNet Top-1 Accuracy73Unverified
3ResNet50ImageNet Top-1 Accuracy71.1Unverified
4ResNet50ImageNet Top-1 Accuracy69.3Unverified
5ResNet50 (v2)ImageNet Top-1 Accuracy67.6Unverified
6ResNet50 (v2)ImageNet Top-1 Accuracy63.8Unverified
7ResNet50ImageNet Top-1 Accuracy63.6Unverified
8ResNet50ImageNet Top-1 Accuracy61.5Unverified
9ResNet50ImageNet Top-1 Accuracy61.5Unverified
10ResNet50 (4×)ImageNet Top-1 Accuracy61.3Unverified
#ModelMetricClaimedVerifiedStatus
110..5sec1Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)84.77Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)85.55Unverified