SOTAVerified

Self-Supervised Learning

Self-Supervised Learning is proposed for utilizing unlabeled data with the success of supervised learning. Producing a dataset with good labels is expensive, while unlabeled data is being generated all the time. The motivation of Self-Supervised Learning is to make use of the large amount of unlabeled data. The main idea of Self-Supervised Learning is to generate the labels from unlabeled data, according to the structure or characteristics of the data itself, and then train on this unsupervised data in a supervised manner. Self-Supervised Learning is wildly used in representation learning to make a model learn the latent features of the data. This technique is often employed in computer vision, video processing and robot control.

Source: Self-supervised Point Set Local Descriptors for Point Cloud Registration

Image source: LeCun

Papers

Showing 31513200 of 5044 papers

TitleStatusHype
Locally Constrained Representations in Reinforcement Learning0
Local Manifold Augmentation for Multiview Semantic Consistency0
Local Policy Improvement for Recommender Systems0
Local-to-Global Self-Supervised Representation Learning for Diabetic Retinopathy Grading0
LOHA: Direct Graph Spectral Contrastive Learning Between Low-pass and High-pass Views0
Longitudinal Self-supervised Learning Using Neural Ordinary Differential Equation0
Longitudinal Self-Supervision for COVID-19 Pathology Quantification0
LoopReg: Self-supervised Learning of Implicit Surface Correspondences, Pose and Shape for 3D Human Mesh Registration0
Low-Data Classification of Historical Music Manuscripts: A Few-Shot Learning Approach0
Low-rank Optimal Transport: Approximation, Statistics and Debiasing0
Low-Res Leads the Way: Improving Generalization for Super-Resolution by Self-Supervised Learning0
Low-Resource Self-Supervised Learning with SSL-Enhanced TTS0
Low-Trace Adaptation of Zero-shot Self-supervised Blind Image Denoising0
LPaintB: Learning to Paint from Self-Supervision0
LSM-2: Learning from Incomplete Wearable Sensor Data0
Longitudinal Self-Supervised Learning0
LSTM Self-Supervision for Detailed Behavior Analysis0
LV-MAE: Learning Long Video Representations through Masked-Embedding Autoencoders0
LW-FedSSL: Resource-efficient Layer-wise Federated Self-supervised Learning0
M2D2: Exploring General-purpose Audio-Language Representations Beyond CLAP0
M2D-CLAP: Masked Modeling Duo Meets CLAP for Learning General-purpose Audio-Language Representation0
M^33D: Learning 3D priors using Multi-Modal Masked Autoencoders for 2D image and video understanding0
MacDiff: Unified Skeleton Modeling with Masked Conditional Diffusion0
Machine Unlearning in Contrastive Learning0
MAEEG: Masked Auto-encoder for EEG Representation Learning0
MAESTRO: Matched Speech Text Representations through Modality Matching0
Making Self-supervised Learning Robust to Spurious Correlation via Learning-speed Aware Sampling0
Mandarin-English Code-switching Speech Recognition with Self-supervised Speech Representation Models0
MAPGN: MAsked Pointer-Generator Network for sequence-to-sequence pre-training0
MAP-Music2Vec: A Simple and Effective Baseline for Self-Supervised Music Audio Representation Learning0
Masked Autoencoder for Unsupervised Video Summarization0
Masked Generative Extractor for Synergistic Representation and 3D Generation of Point Clouds0
Masked Image Modeling Advances 3D Medical Image Analysis0
Masked LoGoNet: Fast and Accurate 3D Image Analysis for Medical Domain0
Masked Modeling Duo for Speech: Specializing General-Purpose Audio Representation to Speech using Denoising Distillation0
Masked Modeling Duo: Learning Representations by Encouraging Both Networks to Model the Input0
Masked Modeling Duo: Towards a Universal Audio Pre-training Framework0
Masked Momentum Contrastive Learning for Zero-shot Semantic Understanding0
Masked Multi-Step Multivariate Time Series Forecasting with Future Information0
Masked prediction tasks: a parameter identifiability view0
Masked Reconstruction Contrastive Learning with Information Bottleneck Principle0
Masked Self-Supervised Pre-Training for Text Recognition Transformers on Large-Scale Datasets0
Masked Self-Supervision for Remaining Useful Lifetime Prediction in Machine Tools0
Masked Siamese ConvNets0
Masked Spatial-Spectral Autoencoders Are Excellent Hyperspectral Defenders0
Mask Hierarchical Features For Self-Supervised Learning0
MaskMatch: Boosting Semi-Supervised Learning Through Mask Autoencoder-Driven Feature Learning0
MaskMol: Knowledge-guided Molecular Image Pre-Training Framework for Activity Cliffs0
MASR: Multi-label Aware Speech Representation0
MAST: Masked Augmentation Subspace Training for Generalizable Self-Supervised Priors0
Show:102550
← PrevPage 64 of 101Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Pretraining: NoneImages & Text57.5Unverified
2Pretraining: ShEDImages & Text54.3Unverified
3Pretraining: e-MixImages & Text48.9Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet50Accuracy91.7Unverified
2ResNet18Accuracy91.02Unverified
3MV-MRAccuracy89.67Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet50average top-1 classification accuracy93.89Unverified
2ResNet18average top-1 classification accuracy92.58Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet50average top-1 classification accuracy72.51Unverified
2ResNet18average top-1 classification accuracy69.31Unverified
#ModelMetricClaimedVerifiedStatus
1CorInfomax (ResNet50)Top-1 Accuracy82.64Unverified
2CorInfomax (ResNet18)Top-1 Accuracy80.48Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet50average top-1 classification accuracy51.84Unverified
2ResNet18average top-1 classification accuracy51.67Unverified
#ModelMetricClaimedVerifiedStatus
1CorInfomax (ResNet18)Top-1 Accuracy93.18Unverified
#ModelMetricClaimedVerifiedStatus
1CorInfomax (ResNet18)Top-1 Accuracy71.61Unverified
#ModelMetricClaimedVerifiedStatus
1Hybrid BYOL-S/CvTAccuracy67.2Unverified
#ModelMetricClaimedVerifiedStatus
1CorInfomax (ResNet50)Top-1 Accuracy54.86Unverified