SOTAVerified

Self-Supervised Image Classification

This is the task of image classification using representations learnt with self-supervised learning. Self-supervised methods generally involve a pretext task that is solved to learn a good representation and a loss function to learn with. One example of a loss function is an autoencoder based loss where the goal is reconstruction of an image pixel-by-pixel. A more popular recent example is a contrastive loss, which measure the similarity of sample pairs in a representation space, and where there can be a varying target instead of a fixed target to reconstruct (as in the case of autoencoders).

A common evaluation protocol is to train a linear classifier on top of (frozen) representations learnt by self-supervised methods. The leaderboards for the linear evaluation protocol can be found below. In practice, it is more common to fine-tune features on a downstream task. An alternative evaluation protocol therefore uses semi-supervised learning and finetunes on a % of the labels. The leaderboards for the finetuning protocol can be accessed here.

You may want to read some blog posts before reading the papers and checking the leaderboards:

There is also Yann LeCun's talk at AAAI-20 which you can watch here (35:00+).

( Image credit: A Simple Framework for Contrastive Learning of Visual Representations )

Papers

Showing 150 of 110 papers

TitleStatusHype
DINOv2: Learning Robust Visual Features without SupervisionCode6
Vision Transformers Need RegistersCode6
Architecture-Agnostic Masked Image Modeling -- From ViT back to CNNCode4
Multi-label Cluster Discrimination for Visual Representation LearningCode4
ONE-PEACE: Exploring One General Representation Model Toward Unlimited ModalitiesCode3
Momentum Contrast for Unsupervised Visual Representation LearningCode3
XCiT: Cross-Covariance Image TransformersCode3
Designing BERT for Convolutional Networks: Sparse and Hierarchical Masked ModelingCode3
Big Self-Supervised Models are Strong Semi-Supervised LearnersCode2
Masked Siamese Networks for Label-Efficient LearningCode2
BEiT: BERT Pre-Training of Image TransformersCode2
Unicom: Universal and Compact Representation Learning for Image RetrievalCode2
Generative Pretraining from PixelsCode2
A Simple Framework for Contrastive Learning of Visual RepresentationsCode2
Stabilize the Latent Space for Image Autoregressive Modeling: A Unified PerspectiveCode2
Unsupervised Learning of Visual Features by Contrasting Cluster AssignmentsCode2
Context Autoencoder for Self-Supervised Representation LearningCode2
ReSSL: Relational Self-Supervised Learning with Weak AugmentationCode1
Bootstrapped Masked Autoencoders for Vision BERT PretrainingCode1
Self-labelling via simultaneous clustering and representation learningCode1
Deep Clustering for Unsupervised Learning of Visual FeaturesCode1
Contrastive Tuning: A Little Help to Make Masked Autoencoders ForgetCode1
Mutual Contrastive Learning for Visual Representation LearningCode1
OBoW: Online Bag-of-Visual-Words Generation for Self-Supervised LearningCode1
ResMLP: Feedforward networks for image classification with data-efficient trainingCode1
Self-Supervised Classification NetworkCode1
Contrastive Multiview CodingCode1
An Empirical Study of Training Self-Supervised Vision TransformersCode1
Bootstrap your own latent: A new approach to self-supervised LearningCode1
Weak Augmentation Guided Relational Self-Supervised LearningCode1
Mugs: A Multi-Granular Self-Supervised Learning FrameworkCode1
iBOT: Image BERT Pre-Training with Online TokenizerCode1
When Do Flat Minima Optimizers Work?Code1
Representation Learning via Invariant Causal MechanismsCode1
CompRess: Self-Supervised Learning by Compressing RepresentationsCode1
Max-Margin Contrastive LearningCode1
Compressive Visual RepresentationsCode1
Barlow Twins: Self-Supervised Learning via Redundancy ReductionCode1
PeCo: Perceptual Codebook for BERT Pre-training of Vision TransformersCode1
Improved Baselines with Momentum Contrastive LearningCode1
Exploring Simple Siamese Representation LearningCode1
MIM-Refiner: A Contrastive Learning Boost from Intermediate Pre-Trained RepresentationsCode1
Contrastive Learning with Stronger AugmentationsCode1
Large Scale Adversarial Representation LearningCode1
Masking meets Supervision: A Strong Learning AllianceCode1
Learning by Sorting: Self-supervised Learning with Group Ordering ConstraintsCode1
Masked Feature Prediction for Self-Supervised Visual Pre-TrainingCode1
Boosting Contrastive Self-Supervised Learning with False Negative CancellationCode1
Masked Autoencoders Are Scalable Vision LearnersCode1
CaCo: Both Positive and Negative Samples are Directly Learnable via Cooperative-adversarial Contrastive LearningCode1
Show:102550
← PrevPage 1 of 3Next →

No leaderboard results yet.