SOTAVerified

Zero-Shot Learning

Zero-shot learning (ZSL) is a model's ability to detect classes never seen during training. The condition is that the classes are not known during supervised learning.

Earlier work in zero-shot learning use attributes in a two-step approach to infer unknown classes. In the computer vision context, more recent advances learn mappings from image feature space to semantic space. Other approaches learn non-linear multimodal embeddings. In the modern NLP context, language models can be evaluated on downstream tasks without fine tuning.

Benchmark datasets for zero-shot learning include aPY, AwA, and CUB, among others.

( Image credit: Prototypical Networks for Few shot Learning in PyTorch )

Further readings:

Papers

Showing 451500 of 1864 papers

TitleStatusHype
Connecting NeRFs, Images, and TextCode0
Progressive Semantic-Guided Vision Transformer for Zero-Shot LearningCode1
Test-Time Adaptation with SaLIP: A Cascade of SAM and CLIP for Zero shot Medical Image SegmentationCode2
Anchor-based Robust Finetuning of Vision-Language Models0
Audio-Visual Generalized Zero-Shot Learning using Pre-Trained Large Multi-Modal ModelsCode1
Condition Monitoring with Incomplete Data: An Integrated Variational Autoencoder and Distance Metric Framework0
High-Discriminative Attribute Feature Learning for Generalized Zero-Shot Learning0
Bootstrapping Chest CT Image Understanding by Distilling Knowledge from X-ray Expert Models0
Forget NLI, Use a Dictionary: Zero-Shot Topic Classification for Low-Resource Languages with Application to LuxembourgishCode0
Label Propagation for Zero-shot Classification with Vision-Language ModelsCode1
Towards Large Language Model driven Reference-less Translation Evaluation for English and Indian Languages0
Emergent Abilities in Reduced-Scale Generative Language ModelsCode0
Diffusion based Zero-shot Medical Image-to-Image Translation for Cross Modality Segmentation0
Training-Free Semantic Segmentation via LLM-Supervision0
Small Language Models Learn Enhanced Reasoning Skills from Medical Textbooks0
X-MIC: Cross-Modal Instance Conditioning for Egocentric Action GeneralizationCode1
VP3D: Unleashing 2D Visual Prompt for Text-to-3D Generation0
VLM-CPL: Consensus Pseudo Labels from Vision-Language Models for Human Annotation-Free Pathological Image ClassificationCode1
Comprehensive Evaluation and Insights into the Use of Large Language Models in the Automation of Behavior-Driven Development Acceptance Test FormulationCode0
Long-CLIP: Unlocking the Long-Text Capability of CLIPCode4
Less but Better: Enabling Generalized Zero-shot Learning Towards Unseen Domains by Intrinsic Learning from Redundant LLM SemanticsCode0
HierCode: A Lightweight Hierarchical Codebook for Zero-shot Chinese Text Recognition0
RAR: Retrieving And Ranking Augmented MLLMs for Visual RecognitionCode2
Just Shift It: Test-Time Prototype Shifting for Zero-Shot Generalization with Vision-Language ModelsCode1
Audio-Visual Compound Expression Recognition Method based on Late Modality Fusion and Rule-based Decision0
Entity6K: A Large Open-Domain Evaluation Dataset for Real-World Entity Recognition0
MEDBind: Unifying Language and Multimodal Medical Data Embeddings0
CLIP-VIS: Adapting CLIP for Open-Vocabulary Video Instance SegmentationCode1
UniBind: LLM-Augmented Unified and Balanced Representation Space to Bind Them All0
Eye-gaze Guided Multi-modal Alignment for Medical Representation LearningCode1
Leveraging Large Language Models to Extract Information on Substance Use Disorder Severity from Clinical Notes: A Zero-shot Learning Approach0
Meta-Prompting for Automating Zero-shot Visual Recognition with LLMsCode1
Boosting Continual Learning of Vision-Language Models via Mixture-of-Experts AdaptersCode3
CoLeCLIP: Open-Domain Continual Learning via Joint Task Prompt and Vocabulary LearningCode1
OpenGraph: Open-Vocabulary Hierarchical 3D Graph Representation in Large-Scale Outdoor EnvironmentsCode2
Select and Distill: Selective Dual-Teacher Knowledge Transfer for Continual Learning on Vision-Language Models0
Sentinel-Guided Zero-Shot Learning: A Collaborative Paradigm without Real Data ExposureCode0
Efficient Prompt Tuning of Large Vision-Language Model for Fine-Grained Ship ClassificationCode0
Segmentation of Knee Bones for Osteoarthritis Assessment: A Comparative Analysis of Supervised, Few-Shot, and Zero-Shot Learning Approaches0
MolBind: Multimodal Alignment of Language, Molecules, and Proteins0
MoralBERT: A Fine-Tuned Language Model for Capturing Moral Values in Social DiscussionsCode1
MENTOR: Multilingual tExt detectioN TOward leaRning by analogy0
Towards Zero-shot Human-Object Interaction Detection via Vision-Language Integration0
Improving deep learning with prior knowledge and cognitive models: A survey on enhancing explainability, adversarial robustness and zero-shot learning0
Zero-Shot ECG Classification with Multimodal Learning and Test-time Clinical Knowledge EnhancementCode2
Split to Merge: Unifying Separated Modalities for Unsupervised Domain AdaptationCode1
Personalized LoRA for Human-Centered Text UnderstandingCode1
CSCNET: Class-Specified Cascaded Network for Compositional Zero-Shot Learning0
MedFLIP: Medical Vision-and-Language Self-supervised Fast Pre-Training with Masked Autoencoder0
ACC-ViT : Atrous Convolution's Comeback in Vision Transformers0
Show:102550
← PrevPage 10 of 38Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ZeroDiffaverage top-1 classification accuracy87.5Unverified
2DUETaverage top-1 classification accuracy72.3Unverified
3Composeraverage top-1 classification accuracy69.4Unverified
4HDC-ZSC-MLPaverage top-1 classification accuracy65.6Unverified
5ZSL_TF-VAEGANaverage top-1 classification accuracy64.9Unverified
6ZLaPAccuracy64.3Unverified
7ZLaP*Accuracy64.2Unverified
8HDC-ZSCaverage top-1 classification accuracy63.8Unverified
9SPOTaverage top-1 classification accuracy62.9Unverified
10f-VAEGAN-D2average top-1 classification accuracy61Unverified
#ModelMetricClaimedVerifiedStatus
1dmis-lab/biobert-v1.1Accuracy26.15Unverified
2meta-llama/Meta-Llama-3-8B-InstructAccuracy25.84Unverified
3epfl-llm/meditron-7bAccuracy25.75Unverified
4dmis-lab/meerkat-7b-v1.0Accuracy25.68Unverified
5meta-llama/Meta-Llama-3-8B-InstructAccuracy25.65Unverified
6HuggingFaceH4/zephyr-7b-betaAccuracy25.54Unverified
7dmis-lab/biobert-v1.1Accuracy25.46Unverified
8epfl-llm/meditron-70bAccuracy25.36Unverified
9epfl-llm/meditron-70bAccuracy25.26Unverified
10HuggingFaceH4/zephyr-7b-betaAccuracy25.06Unverified
#ModelMetricClaimedVerifiedStatus
1ZeroDiffaverage top-1 classification accuracy77.3Unverified
2SPOT (VAEGAN)average top-1 classification accuracy66.04Unverified
3ZSL_TF-VAEGANaverage top-1 classification accuracy66Unverified
4f-VAEGANaverage top-1 classification accuracy64.7Unverified
5DUET (Ours)average top-1 classification accuracy64.4Unverified
6LisGANaverage top-1 classification accuracy61.7Unverified
7TCNaverage top-1 classification accuracy61.5Unverified
8f-CLSWGANaverage top-1 classification accuracy60.8Unverified
9Cycle-WGANaverage top-1 classification accuracy59.9Unverified
#ModelMetricClaimedVerifiedStatus
1ZeroDiffaverage top-1 classification accuracy86.4Unverified
2ZSL-KGaverage top-1 classification accuracy78.08Unverified
3ZSL_TF-VAEGANaverage top-1 classification accuracy72.2Unverified
4DUET (Ours)average top-1 classification accuracy69.9Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy84Unverified
2ZLaP*Accuracy83.1Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy93.6Unverified
2ZLaPAccuracy93.4Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy74.2Unverified
2ZLaPAccuracy74Unverified
#ModelMetricClaimedVerifiedStatus
1ViT-B/16Average mAP60.17Unverified
2ResNet-50Average mAP56.19Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy51.2Unverified
2ZLaP*Accuracy51Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy29.1Unverified
2ZLaP*Accuracy29Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy75.9Unverified
2ZLaP*Accuracy75.5Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy87.9Unverified
2ZLaPAccuracy87.8Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPTop 1 Accuracy72.1Unverified
2ZLaP*Top 1 Accuracy72.1Unverified
#ModelMetricClaimedVerifiedStatus
1HiTeAAccuracy21.7Unverified
2HiTeAAccuracy0.46Unverified
#ModelMetricClaimedVerifiedStatus
1HiTeAAccuracy37.4Unverified
2HiTeAAccuracy0.56Unverified
#ModelMetricClaimedVerifiedStatus
1SPOTaverage top-1 classification accuracy71.9Unverified
2ZSL_TF-VAEGANaverage top-1 classification accuracy70.8Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy90Unverified
2ZLaP*Accuracy89Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy71.8Unverified
2ZLaPAccuracy71.2Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy71.4Unverified
2ZLaPAccuracy71Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy76.3Unverified
2ZLaPAccuracy76.3Unverified
#ModelMetricClaimedVerifiedStatus
1CLIP(ViT-B/16)Average mAP85.77Unverified
2CLIP(ResNet-50)Average mAP84.3Unverified
#ModelMetricClaimedVerifiedStatus
1ZSL-KGTop-160.54Unverified
#ModelMetricClaimedVerifiedStatus
1zsl_ADAAverage Per-Class Accuracy70.9Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy63.2Unverified
#ModelMetricClaimedVerifiedStatus
1MSDAPearson correlation coefficient (PCC)0.52Unverified
#ModelMetricClaimedVerifiedStatus
1SeViLAAccuracy72.3Unverified
#ModelMetricClaimedVerifiedStatus
1M^2-EncoderAccuracy80.7Unverified
#ModelMetricClaimedVerifiedStatus
1FrozenBiLMAccuracy51.5Unverified
#ModelMetricClaimedVerifiedStatus
1CZSLA-acc36Unverified
#ModelMetricClaimedVerifiedStatus
1ZS3Netk=10 mIOU26.3Unverified
#ModelMetricClaimedVerifiedStatus
1ZSL-KGAccuracy88.98Unverified
#ModelMetricClaimedVerifiedStatus
1VideoChat2Accuracy40.6Unverified