SOTAVerified

Self-Supervised Image Classification

This is the task of image classification using representations learnt with self-supervised learning. Self-supervised methods generally involve a pretext task that is solved to learn a good representation and a loss function to learn with. One example of a loss function is an autoencoder based loss where the goal is reconstruction of an image pixel-by-pixel. A more popular recent example is a contrastive loss, which measure the similarity of sample pairs in a representation space, and where there can be a varying target instead of a fixed target to reconstruct (as in the case of autoencoders).

A common evaluation protocol is to train a linear classifier on top of (frozen) representations learnt by self-supervised methods. The leaderboards for the linear evaluation protocol can be found below. In practice, it is more common to fine-tune features on a downstream task. An alternative evaluation protocol therefore uses semi-supervised learning and finetunes on a % of the labels. The leaderboards for the finetuning protocol can be accessed here.

You may want to read some blog posts before reading the papers and checking the leaderboards:

There is also Yann LeCun's talk at AAAI-20 which you can watch here (35:00+).

( Image credit: A Simple Framework for Contrastive Learning of Visual Representations )

Papers

Showing 51100 of 110 papers

TitleStatusHype
Self-Supervised Classification NetworkCode1
Barlow Twins: Self-Supervised Learning via Redundancy ReductionCode1
OBoW: Online Bag-of-Visual-Words Generation for Self-Supervised LearningCode1
Boosting Contrastive Self-Supervised Learning with False Negative CancellationCode1
Exploring Simple Siamese Representation LearningCode1
CompRess: Self-Supervised Learning by Compressing RepresentationsCode1
Representation Learning via Invariant Causal MechanismsCode1
Bootstrap your own latent: A new approach to self-supervised LearningCode1
Prototypical Contrastive Learning of Unsupervised RepresentationsCode1
Improved Baselines with Momentum Contrastive LearningCode1
Self-Supervised Learning of Pretext-Invariant RepresentationsCode1
Self-labelling via simultaneous clustering and representation learningCode1
Large Scale Adversarial Representation LearningCode1
Contrastive Multiview CodingCode1
Deep Clustering for Unsupervised Learning of Visual FeaturesCode1
Representation Learning with Contrastive Predictive CodingCode1
Unsupervised Feature Learning via Non-Parametric Instance DiscriminationCode1
Unsupervised Representation Learning by Predicting Image RotationsCode1
SynCo: Synthetic Hard Negatives in Contrastive Learning for Better Unsupervised Visual RepresentationsCode0
Unsupervised Representation Learning by Balanced Self Attention MatchingCode0
Estimating Physical Information Consistency of Channel Data Augmentation for Remote Sensing Images0
IPCL: Iterative Pseudo-Supervised Contrastive Learning to Improve Self-Supervised Feature RepresentationCode0
Perceptual Group Tokenizer: Building Perception with Iterative Grouping0
Masked Image Residual Learning for Scaling Deeper Vision TransformersCode0
DINO-CXR: A self supervised method based on vision transformer for chest X-ray classification0
MV-MR: multi-views and multi-representations for self-supervised learning and knowledge distillationCode0
All4One: Symbiotic Neighbour Contrastive Learning via Self-Attention and Redundancy ReductionCode0
Improving Visual Representation Learning through Perceptual UnderstandingCode0
Masked Reconstruction Contrastive Learning with Information Bottleneck Principle0
EVA: Exploring the Limits of Masked Visual Representation Learning at ScaleCode0
Exploring Target Representations for Masked AutoencodersCode0
BEiT v2: Masked Image Modeling with Vector-Quantized Visual TokenizersCode0
Model-Aware Contrastive Learning: Towards Escaping the DilemmasCode0
Unsupervised Visual Representation Learning by Synchronous Momentum GroupingCode0
Vision Models Are More Robust And Fair When Pretrained On Uncurated Images Without SupervisionCode0
OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning FrameworkCode0
Pushing the limits of self-supervised ResNets: Can we outperform supervised learning without labels on ImageNet?Code0
Efficient Self-supervised Vision Transformers for Representation LearningCode0
Large-Scale Unsupervised Person Re-Identification with Contrastive Learning0
Divide and Contrast: Self-supervised Learning from Uncurated Data0
With a Little Help from My Friends: Nearest-Neighbor Contrastive Learning of Visual RepresentationsCode0
Self-supervised Pretraining of Visual Features in the WildCode0
Self-supervised Pre-training with Hard Examples Improves Visual Representations0
A Pseudo-labelling Auto-Encoder for unsupervised image classification0
Seed the Views: Hierarchical Semantic Alignment for Contrastive Representation Learning0
A comparative study of semi- and self-supervised semantic segmentation of biomedical microscopy data0
Consensus Clustering With Unsupervised Representation Learning0
What Makes for Good Views for Contrastive Learning?Code0
On Mutual Information Maximization for Representation LearningCode0
Learning Representations by Maximizing Mutual Information Across ViewsCode0
Show:102550
← PrevPage 2 of 3Next →

No leaderboard results yet.