SOTAVerified

zero-shot-classification

Papers

Showing 51100 of 422 papers

TitleStatusHype
Episode-based Prototype Generating Network for Zero-Shot LearningCode1
ProtoCLIP: Prototypical Contrastive Language Image PretrainingCode1
Lite-Mind: Towards Efficient and Robust Brain Representation NetworkCode1
RaVL: Discovering and Mitigating Spurious Correlations in Fine-Tuned Vision-Language ModelsCode1
CLIP meets DINO for Tuning Zero-Shot Classifier using Unlabeled Image CollectionsCode1
Learning Customized Visual Models with Retrieval-Augmented KnowledgeCode1
Beyond the Next Token: Towards Prompt-Robust Zero-Shot Classification via Efficient Multi-Token PredictionCode1
Mind the Gap: Understanding the Modality Gap in Multi-modal Contrastive Representation LearningCode1
ArcheType: A Novel Framework for Open-Source Column Type Annotation using Large Language ModelsCode1
Label Propagation for Zero-shot Classification with Vision-Language ModelsCode1
Advancing Medical Representation Learning Through High-Quality DataCode1
The Parrot Dilemma: Human-Labeled vs. LLM-augmented Data in Classification TasksCode1
Knowledge-enhanced Visual-Language Pretraining for Computational PathologyCode1
Improved Zero-Shot Classification by Adapting VLMs with Text DescriptionsCode1
Improved Probabilistic Image-Text RepresentationsCode1
Interpreting and Analysing CLIP's Zero-Shot Image Classification via Mutual KnowledgeCode1
Latent Embedding Feedback and Discriminative Features for Zero-Shot ClassificationCode1
MineralImage5k: A benchmark for zero-shot raw mineral visual recognition and descriptionCode1
From Local Details to Global Context: Advancing Vision-Language Models with Attention-Based SelectionCode1
AgriCLIP: Adapting CLIP for Agriculture and Livestock via Domain-Specialized Cross-Model AlignmentCode1
CHiLS: Zero-Shot Image Classification with Hierarchical Label SetsCode1
Adversarial Illusions in Multi-Modal EmbeddingsCode1
Exploring the Spectrum of Visio-Linguistic Compositionality and RecognitionCode1
EmoCLIP: A Vision-Language Method for Zero-Shot Video Facial Expression RecognitionCode1
Exploring Vision-Language Models for Imbalanced LearningCode1
Discovering Human Interactions With Novel Objects via Zero-Shot LearningCode1
Contrastive Language-Image Pre-training for the Italian LanguageCode1
Florence: A New Foundation Model for Computer VisionCode1
Attentive Mask CLIPCode1
CLIPArTT: Adaptation of CLIP to New Domains at Test TimeCode1
Zero-Shot Semantic SegmentationCode1
CLIP-DIY: CLIP Dense Inference Yields Open-Vocabulary Semantic Segmentation For-FreeCode1
Discriminative Region-based Multi-Label Zero-Shot LearningCode1
CLIPer: Hierarchically Improving Spatial Representation of CLIP for Open-Vocabulary Semantic SegmentationCode1
CLIP-Guided Source-Free Object Detection in Aerial ImagesCode1
CLIP-Lite: Information Efficient Visual Representation Learning with Language SupervisionCode1
Adversarial Robustification via Text-to-Image Diffusion ModelsCode1
CALIP: Zero-Shot Enhancement of CLIP with Parameter-free AttentionCode1
LidarCLIP or: How I Learned to Talk to Point CloudsCode1
CLIP Under the Microscope: A Fine-Grained Analysis of Multi-Object RepresentationCode1
CLIPure: Purification in Latent Space via CLIP for Adversarially Robust Zero-Shot ClassificationCode1
CLIP-VIS: Adapting CLIP for Open-Vocabulary Video Instance SegmentationCode1
Open-Pose 3D Zero-Shot Learning: Benchmark and ChallengesCode1
LR0.FM: Low-Res Benchmark and Improving Robustness for Zero-Shot Classification in Foundation ModelsCode1
DST-Det: Simple Dynamic Self-Training for Open-Vocabulary Object DetectionCode1
MM-Skin: Enhancing Dermatology Vision-Language Model with an Image-Text Dataset Derived from TextbooksCode1
Deep Learning Models for Multilingual Hate Speech DetectionCode1
PerceptionCLIP: Visual Classification by Inferring and Conditioning on ContextsCode1
No Token Left Behind: Explainability-Aided Image Classification and GenerationCode1
Decoupling Zero-Shot Semantic SegmentationCode1
Show:102550
← PrevPage 2 of 9Next →

No leaderboard results yet.