SOTAVerified

Exploring Localization for Self-supervised Fine-grained Contrastive Learning

2021-06-30Code Available0· sign in to hype

Di wu, Siyuan Li, Zelin Zang, Stan Z. Li

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Self-supervised contrastive learning has demonstrated great potential in learning visual representations. Despite their success in various downstream tasks such as image classification and object detection, self-supervised pre-training for fine-grained scenarios is not fully explored. We point out that current contrastive methods are prone to memorizing background/foreground texture and therefore have a limitation in localizing the foreground object. Analysis suggests that learning to extract discriminative texture information and localization are equally crucial for fine-grained self-supervised pre-training. Based on our findings, we introduce cross-view saliency alignment (CVSA), a contrastive learning framework that first crops and swaps saliency regions of images as a novel view generation and then guides the model to localize on foreground objects via a cross-view alignment loss. Extensive experiments on both small- and large-scale fine-grained classification benchmarks show that CVSA significantly improves the learned representation.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
CUB-200-2011BYOL+CVSA (ResNet-50)Accuracy77.1Unverified
FGVC-AircraftBYOL+CVSA (ResNet-50)Accuracy87.27Unverified
NABirdsBYOL+CVSA (ResNet-50)Accuracy79.64Unverified
Stanford CarsBYOL+CVSA (ResNet-50)Accuracy89.76Unverified

Reproductions