SOTAVerified

Self-Supervised Learning

Self-Supervised Learning is proposed for utilizing unlabeled data with the success of supervised learning. Producing a dataset with good labels is expensive, while unlabeled data is being generated all the time. The motivation of Self-Supervised Learning is to make use of the large amount of unlabeled data. The main idea of Self-Supervised Learning is to generate the labels from unlabeled data, according to the structure or characteristics of the data itself, and then train on this unsupervised data in a supervised manner. Self-Supervised Learning is wildly used in representation learning to make a model learn the latent features of the data. This technique is often employed in computer vision, video processing and robot control.

Source: Self-supervised Point Set Local Descriptors for Point Cloud Registration

Image source: LeCun

Papers

Showing 42514300 of 5044 papers

TitleStatusHype
MST: Masked Self-Supervised Transformer for Visual Representation0
MTI-Net: A Multi-Target Speech Intelligibility Prediction Model0
MTLoc: A Confidence-Based Source-Free Domain Adaptation Approach For Indoor Localization0
MTSMAE: Masked Autoencoders for Multivariate Time-Series Forecasting0
Multi-Airport Delay Prediction with Transformers0
MultiCamCows2024 -- A Multi-view Image Dataset for AI-driven Holstein-Friesian Cattle Re-Identification on a Working Farm0
Multi-Domain Self-Supervised Learning0
Multi-encoder nnU-Net outperforms Transformer models with self-supervised pretraining0
Multi-fidelity surrogate modeling for temperature field prediction using deep convolution neural network0
Multi-Granularity Click Confidence Learning via Self-Distillation in Recommendation0
Multi-Label Self-Supervised Learning with Scene Images0
Multi-Level Graph Contrastive Learning0
Multilevel Saliency-Guided Self-Supervised Learning for Image Anomaly Detection0
Multilingual Speech Recognition using Knowledge Transfer across Learning Processes0
Multi-Mask Self-Supervised Learning for Physics-Guided Neural Networks in Highly Accelerated MRI0
UXLA: A Robust Unsupervised Data Augmentation Framework for Zero-Resource Cross-Lingual NLP0
Multi-modal AI for comprehensive breast cancer prognostication0
Multi-modal Cross-domain Self-supervised Pre-training for fMRI and EEG Fusion0
Multimodal Deep Learning for Stroke Prediction and Detection using Retinal Imaging and Clinical Data0
Multi-Modal Emotion Recognition by Text, Speech and Video Using Pretrained Transformers0
Multi-modal Food Recommendation using Clustering and Self-supervised Learning0
Multimodal Generalized Zero Shot Learning for Gleason Grading using Self-Supervised Learning0
MultiModal-Learning for Predicting Molecular Properties: A Framework Based on Image and Graph Structures0
Multi-modal Learning for WebAssembly Reverse Engineering0
Multi-Modal Pre-Training for Automated Speech Recognition0
Multimodal Representation Learning of Cardiovascular Magnetic Resonance Imaging0
Multi-Modal Representation Learning with Text-Driven Soft Masks0
Multimodal Self-Supervised Learning for Medical Image Analysis0
Multimodal self-supervised learning for lesion localization0
Multi-Modal Self-Supervised Semantic Communication0
Multi-Modal Unsupervised Pre-Training for Surgical Operating Room Workflow Analysis0
Multi-modal Vision Pre-training for Medical Image Analysis0
Multi-network Contrastive Learning Based on Global and Local Representations0
Multi-object tracking with self-supervised associating network0
Multi-OCT-SelfNet: Integrating Self-Supervised Learning with Multi-Source Data Fusion for Enhanced Multi-Class Retinal Disease Classification0
Multi-organ Self-supervised Contrastive Learning for Breast Lesion Segmentation0
Multi-Parameter Molecular MRI Quantification using Physics-Informed Self-Supervised Learning0
Multiple Object Tracking with Correlation Learning0
Image Enhanced Rotation Prediction for Self-Supervised Learning0
Multi-Scale Neighborhood Occupancy Masked Autoencoder for Self-Supervised Learning in LiDAR Point Clouds0
Multi-Scale Patch-Based Representation Learning for Image Anomaly Detection and Segmentation0
Superpixel Semantics Representation and Pre-training for Vision-Language Task0
Multi Self-supervised Pre-fine-tuned Transformer Fusion for Better Intelligent Transportation Detection0
Multi-source Few-shot Domain Adaptation0
Multi-Task Self-Supervised Learning for Disfluency Detection0
Multi-task Self-Supervised Learning for Human Activity Detection0
Multi-Task Self-Supervised Learning for Image Segmentation Task0
Multi-Task Self-Supervised Object Detection via Recycling of Bounding Box Annotations0
Multi-Task Self-Supervised Pre-Training for Music Classification0
Multi-task Voice Activated Framework using Self-supervised Learning0
Show:102550
← PrevPage 86 of 101Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Pretraining: NoneImages & Text57.5Unverified
2Pretraining: ShEDImages & Text54.3Unverified
3Pretraining: e-MixImages & Text48.9Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet50Accuracy91.7Unverified
2ResNet18Accuracy91.02Unverified
3MV-MRAccuracy89.67Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet50average top-1 classification accuracy93.89Unverified
2ResNet18average top-1 classification accuracy92.58Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet50average top-1 classification accuracy72.51Unverified
2ResNet18average top-1 classification accuracy69.31Unverified
#ModelMetricClaimedVerifiedStatus
1CorInfomax (ResNet50)Top-1 Accuracy82.64Unverified
2CorInfomax (ResNet18)Top-1 Accuracy80.48Unverified
#ModelMetricClaimedVerifiedStatus
1ResNet50average top-1 classification accuracy51.84Unverified
2ResNet18average top-1 classification accuracy51.67Unverified
#ModelMetricClaimedVerifiedStatus
1CorInfomax (ResNet18)Top-1 Accuracy93.18Unverified
#ModelMetricClaimedVerifiedStatus
1CorInfomax (ResNet18)Top-1 Accuracy71.61Unverified
#ModelMetricClaimedVerifiedStatus
1Hybrid BYOL-S/CvTAccuracy67.2Unverified
#ModelMetricClaimedVerifiedStatus
1CorInfomax (ResNet50)Top-1 Accuracy54.86Unverified