SOTAVerified

Zero-Shot Image Classification

Zero-shot image classification is a technique in computer vision where a model can classify images into categories that were not present during training. This is achieved by leveraging semantic information about the categories, such as textual descriptions or relationships between classes.

Papers

Showing 2650 of 111 papers

TitleStatusHype
Semantic Compositions Enhance Vision-Language Contrastive Learning0
PathGen-1.6M: 1.6 Million Pathology Image-text Pairs Generation through Multi-agent CollaborationCode2
Mitigate the Gap: Investigating Approaches for Improving Cross-Modal Alignment in CLIPCode2
WATT: Weight Average Test-Time Adaptation of CLIPCode2
BaFTA: Backprop-Free Test-Time Adaptation For Zero-Shot Vision-Language Models0
Mind's Eye: Image Recognition by EEG via Multimodal Similarity-Keeping Contrastive LearningCode1
Multimodal Adversarial Defense for Vision-Language Models by Leveraging One-To-Many Relationships0
It's Not a Modality Gap: Characterizing and Addressing the Contrastive Gap0
What Do You See? Enhancing Zero-Shot Image Classification with Multimodal Large Language ModelsCode0
Who's in and who's out? A case study of multimodal CLIP-filtering in DataCompCode0
Multi-method Integration with Confidence-based Weighting for Zero-shot Image Classification0
MoDE: CLIP Data Experts via ClusteringCode0
A Progressive Framework of Vision-language Knowledge Distillation and Alignment for Multilingual Scene0
Sparse Concept Bottleneck Models: Gumbel Tricks in Contrastive LearningCode1
Learn "No" to Say "Yes" Better: Improving Vision-Language Models via NegationsCode1
Bridge the Modality and Capability Gaps in Vision-Language Model Selection0
Can We Talk Models Into Seeing the World Differently?Code1
PromptKD: Unsupervised Prompt Distillation for Vision-Language ModelsCode3
Exploring Low-Resource Medical Image Classification with Weakly Supervised Prompt Learning0
Image-Caption Encoding for Improving Zero-Shot GeneralizationCode0
Segment Any ChangeCode0
CLAMP: Contrastive LAnguage Model Prompt-tuning0
LightCLIP: Learning Multi-Level Interaction for Lightweight Vision-Language Models0
Towards Difficulty-Agnostic Efficient Transfer Learning for Vision-Language ModelsCode0
Efficient Model-Agnostic Multi-Group Equivariant Networks0
Show:102550
← PrevPage 2 of 5Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1OpenClip H/14 (34B)(Laion2B)Top-1 accuracy30.01Unverified
#ModelMetricClaimedVerifiedStatus
1CLIP (ViT B-32)Average Score56.64Unverified
#ModelMetricClaimedVerifiedStatus
1GLIP (Tiny A)Average Score11.4Unverified