SOTAVerified

Contrastive Learning

Contrastive Learning is a deep learning technique for unsupervised representation learning. The goal is to learn a representation of data such that similar instances are close together in the representation space, while dissimilar instances are far apart.

It has been shown to be effective in various computer vision and natural language processing tasks, including image retrieval, zero-shot learning, and cross-modal retrieval. In these tasks, the learned representations can be used as features for downstream tasks such as classification and clustering.

(Image credit: Schroff et al. 2015)

Papers

Showing 64516500 of 6661 papers

TitleStatusHype
Self-Supervised Contrastive Learning for Efficient User Satisfaction Prediction in Conversational Agents0
Contrastive Learning of General-Purpose Audio RepresentationsCode0
Less can be more in contrastive learning0
CLAR: Contrastive Learning of Auditory Representations0
Self-supervised Co-training for Video Representation LearningCode1
What About Inputing Policy in Value Function: Policy Representation and Policy-extended Value Function ApproximatorCode1
Improving Transformation Invariance in Contrastive Representation LearningCode1
MetaCOVID: A Siamese neural network framework with contrastive loss for n-shot diagnosis of COVID-19 patientsCode0
i-Mix: A Domain-Agnostic Strategy for Contrastive Representation LearningCode1
Unsupervised Natural Language Inference via Decoupled Multimodal Contrastive LearningCode0
Representation Learning via Invariant Causal MechanismsCode1
Fully Unsupervised Person Re-identification viaSelective Contrastive Learning0
Masked Contrastive Representation Learning for Reinforcement LearningCode1
Function Contrastive Learning of Transferable Meta-Representations0
Self-Supervised Ranking for Representation Learning0
Viewmaker Networks: Learning Views for Unsupervised Representation LearningCode1
Contrast and Classify: Training Robust VQA ModelsCode1
An Analysis of Robustness of Non-Lipschitz NetworksCode0
MixCo: Mix-up Contrastive Learning for Visual RepresentationCode1
MS^2L: Multi-Task Self-Supervised Learning for Skeleton Based Action RecognitionCode1
MoCo-CXR: MoCo Pretraining Improves Representation and Transferability of Chest X-ray ModelsCode1
Contrastive Rendering for Ultrasound Image Segmentation0
Contrastive Representation Learning: A Framework and Review0
MMGSD: Multi-Modal Gaussian Shape Descriptors for Correspondence Matching in 1D and 2D Deformable Objects0
Contrastive Learning with Hard Negative SamplesCode1
Unsupervised Representation Learning by InvariancePropagationCode1
Representation Learning for Sequence Data with Deep Autoencoding Predictive ComponentsCode1
Support-set bottlenecks for video-text representation learning0
A Contrastive Learning Approach for Training Variational Autoencoder Priors0
CO2: Consistent Contrast for Unsupervised Visual Representation Learning0
A Simple Framework for Uncertainty in Contrastive Learning0
EqCo: Equivalent Rules for Self-supervised Contrastive LearningCode0
Unsupervised Reference-Free Summary Quality Evaluation via Contrastive LearningCode1
Conditional Negative Sampling for Contrastive Learning of Visual RepresentationsCode0
Contrastive Learning of Medical Visual Representations from Paired Images and TextCode1
Hard Negative Mixing for Contrastive LearningCode1
Joint Contrastive Learning with Infinite PossibilitiesCode1
Driver Anomaly Detection: A Dataset and Contrastive Learning ApproachCode1
Towards Effective Context for Meta-Reinforcement Learning: an Approach based on Contrastive LearningCode0
CoKe: Localized Contrastive Learning for Robust Keypoint Detection0
Function Contrastive Learning of Transferable Representations0
What About Taking Policy as Input of Value Function: Policy-extended Value Function Approximator0
DROPS: Deep Retrieval of Physiological Signals via Attribute-specific Clinical Prototypes0
G-SimCLR: Self-Supervised Contrastive Learning with Guided Projection via Pseudo LabellingCode1
In-sample Contrastive Learning and Consistent Attention for Weakly Supervised Object LocalizationCode0
G-SimCLR : Self-Supervised Contrastive Learning with Guided Projection via Pseudo LabellingCode1
Structure Aware Negative Sampling in Knowledge Graphs0
Label-Efficient Multi-Task Segmentation using Contrastive LearningCode1
Feature Distillation With Guided Adversarial Contrastive Learning0
Contrastive ClusteringCode1
Show:102550
← PrevPage 130 of 134Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ResNet50ImageNet Top-1 Accuracy73.6Unverified
2ResNet50ImageNet Top-1 Accuracy73Unverified
3ResNet50ImageNet Top-1 Accuracy71.1Unverified
4ResNet50ImageNet Top-1 Accuracy69.3Unverified
5ResNet50 (v2)ImageNet Top-1 Accuracy67.6Unverified
6ResNet50 (v2)ImageNet Top-1 Accuracy63.8Unverified
7ResNet50ImageNet Top-1 Accuracy63.6Unverified
8ResNet50ImageNet Top-1 Accuracy61.5Unverified
9ResNet50ImageNet Top-1 Accuracy61.5Unverified
10ResNet50 (4×)ImageNet Top-1 Accuracy61.3Unverified
#ModelMetricClaimedVerifiedStatus
110..5sec1Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)84.77Unverified
#ModelMetricClaimedVerifiedStatus
1IPCL (ResNet18)Accuracy (Top-1)85.55Unverified