SOTAVerified

Contrastive Learning

Contrastive Learning is a deep learning technique for unsupervised representation learning. The goal is to learn a representation of data such that similar instances are close together in the representation space, while dissimilar instances are far apart.

It has been shown to be effective in various computer vision and natural language processing tasks, including image retrieval, zero-shot learning, and cross-modal retrieval. In these tasks, the learned representations can be used as features for downstream tasks such as classification and clustering.

(Image credit: Schroff et al. 2015)

Papers

Showing 151200 of 6661 papers

TitleStatusHype
Delving into Inter-Image Invariance for Unsupervised Visual RepresentationsCode2
Denoising as Adaptation: Noise-Space Domain Adaptation for Image RestorationCode2
Extended Agriculture-Vision: An Extension of a Large Aerial Image Dataset for Agricultural Pattern AnalysisCode2
AddressCLIP: Empowering Vision-Language Models for City-wide Image Address LocalizationCode2
Decoding speech perception from non-invasive brain recordingsCode2
Few-Shot Scene Classification of Optical Remote Sensing Images Leveraging Calibrated Pretext TasksCode2
DecisionNCE: Embodied Multimodal Representations via Implicit Preference LearningCode2
A DeNoising FPN With Transformer R-CNN for Tiny Object DetectionCode2
Decoupling Static and Hierarchical Motion Perception for Referring Video SegmentationCode2
A Unified Framework for 3D Scene UnderstandingCode2
Generalized Parametric Contrastive LearningCode2
Generalized Semantic Contrastive Learning via Embedding Side Information for Few-Shot Object DetectionCode2
GestureDiffuCLIP: Gesture Diffusion Model with CLIP LatentsCode2
CrossPoint: Self-Supervised Cross-Modal Contrastive Learning for 3D Point Cloud UnderstandingCode2
DATR: Unsupervised Domain Adaptive Detection Transformer with Dataset-Level Adaptation and Prototypical AlignmentCode2
Gramian Multimodal Representation Learning and AlignmentCode2
HiCMAE: Hierarchical Contrastive Masked Autoencoder for Self-Supervised Audio-Visual Emotion RecognitionCode2
Hybrid Internal Model: Learning Agile Legged Locomotion with Simulated Robot ResponseCode2
Crafting Better Contrastive Views for Siamese Representation LearningCode2
CoST: Contrastive Learning of Disentangled Seasonal-Trend Representations for Time Series ForecastingCode2
Cross-lingual and Multilingual CLIPCode2
Intriguing Properties of Contrastive LossesCode2
LamRA: Large Multimodal Model as Your Advanced Retrieval AssistantCode2
DCdetector: Dual Attention Contrastive Representation Learning for Time Series Anomaly DetectionCode2
Anchored Preference Optimization and Contrastive Revisions: Addressing Underspecification in AlignmentCode2
DetailCLIP: Detail-Oriented CLIP for Fine-Grained TasksCode2
Large-Scale Pre-training for Person Re-identification with Noisy LabelsCode2
Latent Guard: a Safety Framework for Text-to-image GenerationCode2
Contrastive learning of cell state dynamics in response to perturbationsCode2
Learning Vision from Models Rivals Learning Vision from DataCode2
Contrastive learning of Class-agnostic Activation Map for Weakly Supervised Object Localization and Semantic SegmentationCode2
BrainMVP: Multi-modal Vision Pre-training for Brain Image Analysis using Multi-parametric MRICode2
Contrastive Learning for Unpaired Image-to-Image TranslationCode2
Mantis: Lightweight Calibrated Foundation Model for User-Friendly Time Series ClassificationCode2
SoftCoT++: Test-Time Scaling with Soft Chain-of-Thought ReasoningCode2
MCL: Multi-view Enhanced Contrastive Learning for Chest X-ray Report GenerationCode2
Mimic before Reconstruct: Enhancing Masked Autoencoders with Feature MimickingCode2
Contrastive Learning of Asset Embeddings from Financial Time SeriesCode2
Contrastive Learning Rivals Masked Image Modeling in Fine-tuning via Feature DistillationCode2
Content-Based Search for Deep Generative ModelsCode2
CoNT: Contrastive Neural Text GenerationCode2
Contrasting Deepfakes Diffusion via Contrastive Learning and Global-Local SimilaritiesCode2
A Comprehensive Survey on Self-Supervised Learning for RecommendationCode2
Neighborhood-Enhanced Supervised Contrastive Learning for Collaborative FilteringCode2
An Experimental Study on Exploring Strong Lightweight Vision Transformers via Masked Image Modeling Pre-TrainingCode2
One Model to Rule them All: Towards Universal Segmentation for Medical Images with Text PromptsCode2
One Trajectory, One Token: Grounded Video Tokenization via Panoptic Sub-object TrajectoryCode2
OpenShape: Scaling Up 3D Shape Representation Towards Open-World UnderstandingCode2
CLIP-Art: Contrastive Pre-training for Fine-Grained Art ClassificationCode2
Contrastive Audio-Visual Masked AutoencoderCode2
Show:102550
← PrevPage 4 of 134Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50ImageNet Top-1 Accuracy73.6Unverified
2ResNet50ImageNet Top-1 Accuracy73Unverified
3ResNet50ImageNet Top-1 Accuracy71.1Unverified
4ResNet50ImageNet Top-1 Accuracy69.3Unverified
5ResNet50 (v2)ImageNet Top-1 Accuracy67.6Unverified
6ResNet50 (v2)ImageNet Top-1 Accuracy63.8Unverified
7ResNet50ImageNet Top-1 Accuracy63.6Unverified
8ResNet50ImageNet Top-1 Accuracy61.5Unverified
9ResNet50ImageNet Top-1 Accuracy61.5Unverified
10ResNet50 (4×)ImageNet Top-1 Accuracy61.3Unverified
#ModelMetricClaimedVerifiedStatus
110..5sec1Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)84.77Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)85.55Unverified