SOTAVerified

Zero-Shot Learning

Zero-shot learning (ZSL) is a model's ability to detect classes never seen during training. The condition is that the classes are not known during supervised learning.

Earlier work in zero-shot learning use attributes in a two-step approach to infer unknown classes. In the computer vision context, more recent advances learn mappings from image feature space to semantic space. Other approaches learn non-linear multimodal embeddings. In the modern NLP context, language models can be evaluated on downstream tasks without fine tuning.

Benchmark datasets for zero-shot learning include aPY, AwA, and CUB, among others.

( Image credit: Prototypical Networks for Few shot Learning in PyTorch )

Further readings:

Papers

Showing 201250 of 1864 papers

TitleStatusHype
Optimizing Large Language Models for Turkish: New Methodologies in Corpus Selection and Training0
Enhancing Robustness of CLIP to Common Corruptions through Bimodal Test-Time Adaptation0
Multimodal Remote Sensing Scene Classification Using VLMs and Dual-Cross Attention NetworksCode0
The use of large language models to enhance cancer clinical trial educational materials0
Perturb and Recover: Fine-tuning for Effective Backdoor Removal from CLIPCode0
Vision Technologies with Applications in Traffic Surveillance Systems: A Holistic Survey0
Proto Successor Measure: Representing the Space of All Possible Solutions of Reinforcement Learning0
Multimodal Whole Slide Foundation Model for PathologyCode4
CLIP meets DINO for Tuning Zero-Shot Classifier using Unlabeled Image CollectionsCode1
Active Data Curation Effectively Distills Large-Scale Multimodal Models0
FLEX-CLIP: Feature-Level GEneration Network Enhanced CLIP for X-shot Cross-modal Retrieval0
TableTime: Reformulating Time Series Classification as Zero-Shot Table Understanding via Large Language ModelsCode1
ChatBCI: A P300 Speller BCI Leveraging Large Language Models for Improved Sentence Composition in Realistic Scenarios0
CLIPer: Hierarchically Improving Spatial Representation of CLIP for Open-Vocabulary Semantic SegmentationCode1
Leveraging MLLM Embeddings and Attribute Smoothing for Compositional Zero-Shot LearningCode1
Visual-Semantic Graph Matching Net for Zero-Shot LearningCode0
TDSM: Triplet Diffusion for Skeleton-Text Matching in Zero-Shot Action RecognitionCode1
CorrCLIP: Reconstructing Correlations in CLIP with Off-the-Shelf Foundation Models for Open-Vocabulary Semantic SegmentationCode2
Measuring similarity between embedding spaces using induced neighborhood graphs0
NatureLM-audio: an Audio-Language Foundation Model for Bioacoustics0
Emotional Images: Assessing Emotions in Images and Potential Biases in Generative Models0
ZOPP: A Framework of Zero-shot Offboard Panoptic Perception for Autonomous Driving0
Enhancing Visual Classification using Comparative DescriptorsCode0
Asterisk*: Keep it Simple0
RaVL: Discovering and Mitigating Spurious Correlations in Fine-Tuned Vision-Language ModelsCode1
A Mamba Foundation Model for Time Series Forecasting0
Abstracted Shapes as Tokens -- A Generalizable and Interpretable Model for Time-series ClassificationCode1
ResiDual Transformer Alignment with Spectral Decomposition0
Large Language Models for Patient Comments Multi-Label Classification0
SleepNetZero: Zero-Burden Zero-Shot Reliable Sleep Staging With Neural Networks Based on Ballistocardiograms0
Automated Feedback in Math Education: A Comparative Analysis of LLMs for Open-Ended Responses0
Active Learning for Vision-Language Models0
Fine-tuned Large Language Models (LLMs): Improved Prompt Injection Attacks Detection0
Label Set Optimization via Activation Distribution Kurtosis for Zero-shot Classification with Generative Models0
MoRE: Multi-Modal Contrastive Pre-training with Transformers on X-Rays, ECGs, and Diagnostic ReportCode0
Assessing Open-world Forgetting in Generative Image Model Customization0
Can Medical Vision-Language Pre-training Succeed with Purely Synthetic Data?0
Interpreting and Analysing CLIP's Zero-Shot Image Classification via Mutual KnowledgeCode1
LLM Chain Ensembles for Scalable and Accurate Data AnnotationCode0
Towards Graph Foundation Models: Training on Knowledge Graphs Enables Transferability to General Graphs0
CtrlSynth: Controllable Image Text Synthesis for Data-Efficient Multimodal Learning0
PSVMA+: Exploring Multi-granularity Semantic-visual Adaption for Generalized Zero-shot Learning0
Continual Learning Improves Zero-Shot Action Recognition0
Large Model for Small Data: Foundation Model for Cross-Modal RF Human Activity Recognition0
Retrieval Instead of Fine-tuning: A Retrieval-based Parameter Ensemble for Zero-shot Learning0
A Unified Debiasing Approach for Vision-Language Models across Modalities and TasksCode0
Collusion Detection with Graph Neural Networks0
Zero-Shot Learning of Causal Models0
ClaimBrush: A Novel Framework for Automated Patent Claim Refinement Based on Large Language Models0
GLOV: Guided Large Language Models as Implicit Optimizers for Vision Language ModelsCode0
Show:102550
← PrevPage 5 of 38Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1ZeroDiffaverage top-1 classification accuracy87.5Unverified
2DUETaverage top-1 classification accuracy72.3Unverified
3Composeraverage top-1 classification accuracy69.4Unverified
4HDC-ZSC-MLPaverage top-1 classification accuracy65.6Unverified
5ZSL_TF-VAEGANaverage top-1 classification accuracy64.9Unverified
6ZLaPAccuracy64.3Unverified
7ZLaP*Accuracy64.2Unverified
8HDC-ZSCaverage top-1 classification accuracy63.8Unverified
9SPOTaverage top-1 classification accuracy62.9Unverified
10f-VAEGAN-D2average top-1 classification accuracy61Unverified
#ModelMetricClaimedVerifiedStatus
1dmis-lab/biobert-v1.1Accuracy26.15Unverified
2meta-llama/Meta-Llama-3-8B-InstructAccuracy25.84Unverified
3epfl-llm/meditron-7bAccuracy25.75Unverified
4dmis-lab/meerkat-7b-v1.0Accuracy25.68Unverified
5meta-llama/Meta-Llama-3-8B-InstructAccuracy25.65Unverified
6HuggingFaceH4/zephyr-7b-betaAccuracy25.54Unverified
7dmis-lab/biobert-v1.1Accuracy25.46Unverified
8epfl-llm/meditron-70bAccuracy25.36Unverified
9epfl-llm/meditron-70bAccuracy25.26Unverified
10HuggingFaceH4/zephyr-7b-betaAccuracy25.06Unverified
#ModelMetricClaimedVerifiedStatus
1ZeroDiffaverage top-1 classification accuracy77.3Unverified
2SPOT (VAEGAN)average top-1 classification accuracy66.04Unverified
3ZSL_TF-VAEGANaverage top-1 classification accuracy66Unverified
4f-VAEGANaverage top-1 classification accuracy64.7Unverified
5DUET (Ours)average top-1 classification accuracy64.4Unverified
6LisGANaverage top-1 classification accuracy61.7Unverified
7TCNaverage top-1 classification accuracy61.5Unverified
8f-CLSWGANaverage top-1 classification accuracy60.8Unverified
9Cycle-WGANaverage top-1 classification accuracy59.9Unverified
#ModelMetricClaimedVerifiedStatus
1ZeroDiffaverage top-1 classification accuracy86.4Unverified
2ZSL-KGaverage top-1 classification accuracy78.08Unverified
3ZSL_TF-VAEGANaverage top-1 classification accuracy72.2Unverified
4DUET (Ours)average top-1 classification accuracy69.9Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy84Unverified
2ZLaP*Accuracy83.1Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy93.6Unverified
2ZLaPAccuracy93.4Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy74.2Unverified
2ZLaPAccuracy74Unverified
#ModelMetricClaimedVerifiedStatus
1ViT-B/16Average mAP60.17Unverified
2ResNet-50Average mAP56.19Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy51.2Unverified
2ZLaP*Accuracy51Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy29.1Unverified
2ZLaP*Accuracy29Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy75.9Unverified
2ZLaP*Accuracy75.5Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy87.9Unverified
2ZLaPAccuracy87.8Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPTop 1 Accuracy72.1Unverified
2ZLaP*Top 1 Accuracy72.1Unverified
#ModelMetricClaimedVerifiedStatus
1HiTeAAccuracy21.7Unverified
2HiTeAAccuracy0.46Unverified
#ModelMetricClaimedVerifiedStatus
1HiTeAAccuracy37.4Unverified
2HiTeAAccuracy0.56Unverified
#ModelMetricClaimedVerifiedStatus
1SPOTaverage top-1 classification accuracy71.9Unverified
2ZSL_TF-VAEGANaverage top-1 classification accuracy70.8Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaPAccuracy90Unverified
2ZLaP*Accuracy89Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy71.8Unverified
2ZLaPAccuracy71.2Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy71.4Unverified
2ZLaPAccuracy71Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy76.3Unverified
2ZLaPAccuracy76.3Unverified
#ModelMetricClaimedVerifiedStatus
1CLIP(ViT-B/16)Average mAP85.77Unverified
2CLIP(ResNet-50)Average mAP84.3Unverified
#ModelMetricClaimedVerifiedStatus
1ZSL-KGTop-160.54Unverified
#ModelMetricClaimedVerifiedStatus
1zsl_ADAAverage Per-Class Accuracy70.9Unverified
#ModelMetricClaimedVerifiedStatus
1ZLaP*Accuracy63.2Unverified
#ModelMetricClaimedVerifiedStatus
1MSDAPearson correlation coefficient (PCC)0.52Unverified
#ModelMetricClaimedVerifiedStatus
1SeViLAAccuracy72.3Unverified
#ModelMetricClaimedVerifiedStatus
1M^2-EncoderAccuracy80.7Unverified
#ModelMetricClaimedVerifiedStatus
1FrozenBiLMAccuracy51.5Unverified
#ModelMetricClaimedVerifiedStatus
1CZSLA-acc36Unverified
#ModelMetricClaimedVerifiedStatus
1ZS3Netk=10 mIOU26.3Unverified
#ModelMetricClaimedVerifiedStatus
1ZSL-KGAccuracy88.98Unverified
#ModelMetricClaimedVerifiedStatus
1VideoChat2Accuracy40.6Unverified