SOTAVerified

Self-Supervised Image Classification

This is the task of image classification using representations learnt with self-supervised learning. Self-supervised methods generally involve a pretext task that is solved to learn a good representation and a loss function to learn with. One example of a loss function is an autoencoder based loss where the goal is reconstruction of an image pixel-by-pixel. A more popular recent example is a contrastive loss, which measure the similarity of sample pairs in a representation space, and where there can be a varying target instead of a fixed target to reconstruct (as in the case of autoencoders).

A common evaluation protocol is to train a linear classifier on top of (frozen) representations learnt by self-supervised methods. The leaderboards for the linear evaluation protocol can be found below. In practice, it is more common to fine-tune features on a downstream task. An alternative evaluation protocol therefore uses semi-supervised learning and finetunes on a % of the labels. The leaderboards for the finetuning protocol can be accessed here.

You may want to read some blog posts before reading the papers and checking the leaderboards:

There is also Yann LeCun's talk at AAAI-20 which you can watch here (35:00+).

( Image credit: A Simple Framework for Contrastive Learning of Visual Representations )

Papers

Showing 2650 of 110 papers

TitleStatusHype
Exploring Target Representations for Masked AutoencodersCode0
BEiT v2: Masked Image Modeling with Vector-Quantized Visual TokenizersCode0
Model-Aware Contrastive Learning: Towards Escaping the DilemmasCode0
Bootstrapped Masked Autoencoders for Vision BERT PretrainingCode1
Unsupervised Visual Representation Learning by Synchronous Momentum GroupingCode0
Architecture-Agnostic Masked Image Modeling -- From ViT back to CNNCode4
Multiplexed Immunofluorescence Brain Image Analysis Using Self-Supervised Dual-Loss Adaptive Masked AutoencoderCode1
Masked Siamese Networks for Label-Efficient LearningCode2
mc-BEiT: Multi-choice Discretization for Image BERT Pre-trainingCode1
Mugs: A Multi-Granular Self-Supervised Learning FrameworkCode1
CaCo: Both Positive and Negative Samples are Directly Learnable via Cooperative-adversarial Contrastive LearningCode1
Weak Augmentation Guided Relational Self-Supervised LearningCode1
Vision Models Are More Robust And Fair When Pretrained On Uncurated Images Without SupervisionCode0
Context Autoencoder for Self-Supervised Representation LearningCode2
OFA: Unifying Architectures, Tasks, and Modalities Through a Simple Sequence-to-Sequence Learning FrameworkCode0
When Do Flat Minima Optimizers Work?Code1
Pushing the limits of self-supervised ResNets: Can we outperform supervised learning without labels on ImageNet?Code0
Max-Margin Contrastive LearningCode1
Masked Feature Prediction for Self-Supervised Visual Pre-TrainingCode1
Similarity Contrastive Estimation for Self-Supervised Soft Contrastive LearningCode1
PeCo: Perceptual Codebook for BERT Pre-training of Vision TransformersCode1
SimMIM: A Simple Framework for Masked Image ModelingCode1
iBOT: Image BERT Pre-Training with Online TokenizerCode1
Masked Autoencoders Are Scalable Vision LearnersCode1
Self-Supervised Learning by Estimating Twin Class DistributionsCode1
Show:102550
← PrevPage 2 of 5Next →

No leaderboard results yet.