SOTAVerified

Self-Supervised Image Classification

This is the task of image classification using representations learnt with self-supervised learning. Self-supervised methods generally involve a pretext task that is solved to learn a good representation and a loss function to learn with. One example of a loss function is an autoencoder based loss where the goal is reconstruction of an image pixel-by-pixel. A more popular recent example is a contrastive loss, which measure the similarity of sample pairs in a representation space, and where there can be a varying target instead of a fixed target to reconstruct (as in the case of autoencoders).

A common evaluation protocol is to train a linear classifier on top of (frozen) representations learnt by self-supervised methods. The leaderboards for the linear evaluation protocol can be found below. In practice, it is more common to fine-tune features on a downstream task. An alternative evaluation protocol therefore uses semi-supervised learning and finetunes on a % of the labels. The leaderboards for the finetuning protocol can be accessed here.

You may want to read some blog posts before reading the papers and checking the leaderboards:

There is also Yann LeCun's talk at AAAI-20 which you can watch here (35:00+).

( Image credit: A Simple Framework for Contrastive Learning of Visual Representations )

Papers

Showing 7180 of 110 papers

TitleStatusHype
Unsupervised Visual Representation Learning by Synchronous Momentum Grouping0
Estimating Physical Information Consistency of Channel Data Augmentation for Remote Sensing Images0
Perceptual Group Tokenizer: Building Perception with Iterative Grouping0
Seed the Views: Hierarchical Semantic Alignment for Contrastive Representation Learning0
A Pseudo-labelling Auto-Encoder for unsupervised image classification0
What Makes for Good Views for Contrastive Learning?0
Large-Scale Unsupervised Person Re-Identification with Contrastive Learning0
Vision Models Are More Robust And Fair When Pretrained On Uncurated Images Without Supervision0
Masked Reconstruction Contrastive Learning with Information Bottleneck Principle0
DINO-CXR: A self supervised method based on vision transformer for chest X-ray classification0
Show:102550
← PrevPage 8 of 11Next →

No leaderboard results yet.