SOTAVerified

Zero-Shot Learning

Zero-shot learning (ZSL) is a model's ability to detect classes never seen during training. The condition is that the classes are not known during supervised learning.

Earlier work in zero-shot learning use attributes in a two-step approach to infer unknown classes. In the computer vision context, more recent advances learn mappings from image feature space to semantic space. Other approaches learn non-linear multimodal embeddings. In the modern NLP context, language models can be evaluated on downstream tasks without fine tuning.

Benchmark datasets for zero-shot learning include aPY, AwA, and CUB, among others.

( Image credit: Prototypical Networks for Few shot Learning in PyTorch )

Further readings:

Papers

Showing 401450 of 1864 papers

TitleStatusHype
DarijaBanking: A New Resource for Overcoming Language Barriers in Banking Intent Detection for Moroccan Arabic SpeakersCode0
CapS-Adapter: Caption-based MultiModal Adapter in Zero-Shot ClassificationCode0
BDetCLIP: Multimodal Prompting Contrastive Test-Time Backdoor Detection0
What Do You See? Enhancing Zero-Shot Image Classification with Multimodal Large Language ModelsCode0
CLIP model is an Efficient Online Lifelong LearnerCode0
Chain-of-Thought Prompting for Demographic Inference with Large Multimodal ModelsCode0
Benchmarking the Performance of Pre-trained LLMs across Urdu NLP Tasks0
Magnetic Resonance Image Processing Transformer for General Accelerated Image Reconstruction0
Tuning-free Universally-Supervised Semantic Segmentation0
Harmony: A Joint Self-Supervised and Weakly-Supervised Framework for Learning General Purpose Visual RepresentationsCode0
Implicit In-context LearningCode1
Floor-Plan-aided Indoor Localization: Zero-Shot Learning Framework, Data Sets, and PrototypeCode0
Evaluating and Modeling Social Intelligence: A Comparative Study of Human and AI CapabilitiesCode0
LiPost: Improved Content Understanding With Effective Use of Multi-task Contrastive Learning0
Stylometric Watermarks for Large Language Models0
CLIP with Quality Captions: A Strong Pretraining for Vision Tasks0
MedConceptsQA: Open Source Medical Concepts QA BenchmarkCode1
Differentiable Model Scaling using Differentiable TopkCode1
Advanced Natural-based interaction for the ITAlian language: LLaMAntino-3-ANITA0
Pseudo-Prompt Generating in Pre-trained Vision-Language Models for Multi-Label Medical Image ClassificationCode1
The Effect of Model Size on LLM Post-hoc Explainability via LIMECode0
Enhancing Q-Learning with Large Language Model Heuristics0
Dual Relation Mining Network for Zero-Shot Learning0
Adapting Dual-encoder Vision-language Models for Paraphrased Retrieval0
CICA: Content-Injected Contrastive Alignment for Zero-Shot Document Image Classification0
Source-Free Domain Adaptation Guided by Vision and Vision-Language Pre-TrainingCode0
Multi-method Integration with Confidence-based Weighting for Zero-shot Image Classification0
Exploiting ChatGPT for Diagnosing Autism-Associated Language Disorders and Identifying Distinct FeaturesCode0
CLIPArTT: Adaptation of CLIP to New Domains at Test TimeCode1
Modeling Caption Diversity in Contrastive Vision-Language PretrainingCode1
PEVA-Net: Prompt-Enhanced View Aggregation Network for Zero/Few-Shot Multi-View 3D Shape Recognition0
Weighted Point Cloud Embedding for Multimodal Contrastive Learning Toward Optimal Similarity Metric0
CLIP-Mamba: CLIP Pretrained Mamba Models with OOD and Hessian EvaluationCode2
ESP-Zero: Unsupervised enhancement of zero-shot classification for Extremely Sparse Point cloud0
Prevalent Frequency of Emotional and Physical Symptoms in Social Anxiety using Zero Shot Classification: An Observational Study0
Dual Expert Distillation Network for Generalized Zero-Shot LearningCode0
OpenDlign: Open-World Point Cloud Understanding with Depth-Aligned ImagesCode1
Zero-Shot Distillation for Image Encoders: How to Make Effective Use of Synthetic Data0
AAPL: Adding Attributes to Prompt Learning for Vision-Language ModelsCode1
Embracing Diversity: Interpretable Zero-shot classification beyond one vector per class0
Visual-Augmented Dynamic Semantic Prototype for Generative Zero-Shot Learning0
Small Language Models are Good Too: An Empirical Study of Zero-Shot Classification0
Two-Stage Stance Labeling: User-Hashtag Heuristics with Graph Neural Networks0
Knowledge-enhanced Visual-Language Pretraining for Computational PathologyCode1
Evolving Interpretable Visual Classifiers with Large Language Models0
The Devil is in the Few Shots: Iterative Visual Knowledge Completion for Few-shot LearningCode1
CREST: Cross-modal Resonance through Evidential Deep Learning for Enhanced Zero-Shot LearningCode0
Customising General Large Language Models for Specialised Emotion Recognition TasksCode0
`Eyes of a Hawk and Ears of a Fox': Part Prototype Network for Generalized Zero-Shot Learning0
OTTER: Effortless Label Distribution Adaptation of Zero-shot ModelsCode0
Show:102550
← PrevPage 9 of 38Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ZeroDiffaverage top-1 classification accuracy87.5Unverified
2DUETaverage top-1 classification accuracy72.3Unverified
3Composeraverage top-1 classification accuracy69.4Unverified
4HDC-ZSC-MLPaverage top-1 classification accuracy65.6Unverified
5ZSL_TF-VAEGANaverage top-1 classification accuracy64.9Unverified
6ZLaPAccuracy64.3Unverified
7ZLaP*Accuracy64.2Unverified
8HDC-ZSCaverage top-1 classification accuracy63.8Unverified
9SPOTaverage top-1 classification accuracy62.9Unverified
10f-VAEGAN-D2average top-1 classification accuracy61Unverified
#ModelMetricClaimedVerifiedStatus
1dmis-lab/biobert-v1.1Accuracy26.15Unverified
2meta-llama/Meta-Llama-3-8B-InstructAccuracy25.84Unverified
3epfl-llm/meditron-7bAccuracy25.75Unverified
4dmis-lab/meerkat-7b-v1.0Accuracy25.68Unverified
5meta-llama/Meta-Llama-3-8B-InstructAccuracy25.65Unverified
6HuggingFaceH4/zephyr-7b-betaAccuracy25.54Unverified
7dmis-lab/biobert-v1.1Accuracy25.46Unverified
8epfl-llm/meditron-70bAccuracy25.36Unverified
9epfl-llm/meditron-70bAccuracy25.26Unverified
10HuggingFaceH4/zephyr-7b-betaAccuracy25.06Unverified
#ModelMetricClaimedVerifiedStatus
1ZeroDiffaverage top-1 classification accuracy77.3Unverified
2SPOT (VAEGAN)average top-1 classification accuracy66.04Unverified
3ZSL_TF-VAEGANaverage top-1 classification accuracy66Unverified
4f-VAEGANaverage top-1 classification accuracy64.7Unverified
5DUET (Ours)average top-1 classification accuracy64.4Unverified
6LisGANaverage top-1 classification accuracy61.7Unverified
7TCNaverage top-1 classification accuracy61.5Unverified
8f-CLSWGANaverage top-1 classification accuracy60.8Unverified
9Cycle-WGANaverage top-1 classification accuracy59.9Unverified
#ModelMetricClaimedVerifiedStatus
1ZeroDiffaverage top-1 classification accuracy86.4Unverified
2ZSL-KGaverage top-1 classification accuracy78.08Unverified
3ZSL_TF-VAEGANaverage top-1 classification accuracy72.2Unverified
4DUET (Ours)average top-1 classification accuracy69.9Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy84Unverified
2ZLaP*Accuracy83.1Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy93.6Unverified
2ZLaPAccuracy93.4Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy74.2Unverified
2ZLaPAccuracy74Unverified
#ModelMetricClaimedVerifiedStatus
1ViT-B/16Average mAP60.17Unverified
2ResNet-50Average mAP56.19Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy51.2Unverified
2ZLaP*Accuracy51Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy29.1Unverified
2ZLaP*Accuracy29Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy75.9Unverified
2ZLaP*Accuracy75.5Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy87.9Unverified
2ZLaPAccuracy87.8Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPTop 1 Accuracy72.1Unverified
2ZLaP*Top 1 Accuracy72.1Unverified
#ModelMetricClaimedVerifiedStatus
1HiTeAAccuracy21.7Unverified
2HiTeAAccuracy0.46Unverified
#ModelMetricClaimedVerifiedStatus
1HiTeAAccuracy37.4Unverified
2HiTeAAccuracy0.56Unverified
#ModelMetricClaimedVerifiedStatus
1SPOTaverage top-1 classification accuracy71.9Unverified
2ZSL_TF-VAEGANaverage top-1 classification accuracy70.8Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy90Unverified
2ZLaP*Accuracy89Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy71.8Unverified
2ZLaPAccuracy71.2Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy71.4Unverified
2ZLaPAccuracy71Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy76.3Unverified
2ZLaP*Accuracy76.3Unverified
#ModelMetricClaimedVerifiedStatus
1CLIP(ViT-B/16)Average mAP85.77Unverified
2CLIP(ResNet-50)Average mAP84.3Unverified
#ModelMetricClaimedVerifiedStatus
1ZSL-KGTop-160.54Unverified
#ModelMetricClaimedVerifiedStatus
1zsl_ADAAverage Per-Class Accuracy70.9Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy63.2Unverified
#ModelMetricClaimedVerifiedStatus
1MSDAPearson correlation coefficient (PCC)0.52Unverified
#ModelMetricClaimedVerifiedStatus
1SeViLAAccuracy72.3Unverified
#ModelMetricClaimedVerifiedStatus
1M^2-EncoderAccuracy80.7Unverified
#ModelMetricClaimedVerifiedStatus
1FrozenBiLMAccuracy51.5Unverified
#ModelMetricClaimedVerifiedStatus
1CZSLA-acc36Unverified
#ModelMetricClaimedVerifiedStatus
1ZS3Netk=10 mIOU26.3Unverified
#ModelMetricClaimedVerifiedStatus
1ZSL-KGAccuracy88.98Unverified
#ModelMetricClaimedVerifiedStatus
1VideoChat2Accuracy40.6Unverified