SOTAVerified

Contrastive Learning

Contrastive Learning is a deep learning technique for unsupervised representation learning. The goal is to learn a representation of data such that similar instances are close together in the representation space, while dissimilar instances are far apart.

It has been shown to be effective in various computer vision and natural language processing tasks, including image retrieval, zero-shot learning, and cross-modal retrieval. In these tasks, the learned representations can be used as features for downstream tasks such as classification and clustering.

(Image credit: Schroff et al. 2015)

Papers

Showing 376400 of 6661 papers

TitleStatusHype
NaFM: Pre-training a Foundation Model for Small-Molecule Natural ProductsCode0
Neural-Guided Equation DiscoveryCode0
Should we pre-train a decoder in contrastive learning for dense prediction tasks?0
Generative Modeling of Class Probability for Multi-Modal Representation Learning0
Semi-supervised Cervical Segmentation on Ultrasound by A Dual Framework for Neural NetworksCode0
A Statistical Theory of Contrastive Learning via Approximate Sufficient Statistics0
Federated Cross-Domain Click-Through Rate Prediction With Large Language Model Augmentation0
GAIR: Improving Multimodal Geo-Foundation Model with Geo-Aligned Implicit Representations0
DocVideoQA: Towards Comprehensive Understanding of Document-Centric Videos through Question Answering0
Diffusion-augmented Graph Contrastive Learning for Collaborative Filter0
Beyond the Visible: Multispectral Vision-Language Learning for Earth Observation0
UniSync: A Unified Framework for Audio-Visual Synchronization0
PromptHash: Affinity-Prompted Collaborative Cross-Modal Learning for Adaptive Hashing RetrievalCode0
Unlocking the Capabilities of Vision-Language Models for Generalizable and Explainable Deepfake Detection0
Continual Multimodal Contrastive Learning0
TULIP: Towards Unified Language-Image Pretraining0
Graph-Weighted Contrastive Learning for Semi-Supervised Hyperspectral Image ClassificationCode0
EgoDTM: Towards 3D-Aware Egocentric Video-Language PretrainingCode0
Machine Unlearning in Hyperbolic vs. Euclidean Multimodal Contrastive Learning: Adapting Alignment Calibration to MERUCode0
Text-Derived Relational Graph-Enhanced Network for Skeleton-Based Action SegmentationCode0
DashCLIP: Leveraging multimodal models for generating semantic embeddings for DoorDash0
EEG-CLIP : Learning EEG representations from natural language descriptionsCode1
Incorporating Attributes and Multi-Scale Structures for Heterogeneous Graph Contrastive Learning0
Unlocking the Potential of Unlabeled Data in Semi-Supervised Domain GeneralizationCode1
DualToken: Towards Unifying Visual Understanding and Generation with Dual Visual Vocabularies0
Show:102550
← PrevPage 16 of 267Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50ImageNet Top-1 Accuracy73.6Unverified
2ResNet50ImageNet Top-1 Accuracy73Unverified
3ResNet50ImageNet Top-1 Accuracy71.1Unverified
4ResNet50ImageNet Top-1 Accuracy69.3Unverified
5ResNet50 (v2)ImageNet Top-1 Accuracy67.6Unverified
6ResNet50 (v2)ImageNet Top-1 Accuracy63.8Unverified
7ResNet50ImageNet Top-1 Accuracy63.6Unverified
8ResNet50ImageNet Top-1 Accuracy61.5Unverified
9ResNet50ImageNet Top-1 Accuracy61.5Unverified
10ResNet50 (4×)ImageNet Top-1 Accuracy61.3Unverified
#ModelMetricClaimedVerifiedStatus
110..5sec1Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)84.77Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)85.55Unverified