SOTAVerified

Open Vocabulary Semantic Segmentation

Papers

Showing 51100 of 113 papers

TitleStatusHype
SegPoint: Segment Any Point Cloud via Large Language Model0
ClearCLIP: Decomposing CLIP Representations for Dense Vision-Language Inference0
Explore the Potential of CLIP for Training-Free Open Vocabulary Semantic SegmentationCode1
Test-time Contrastive Concepts for Open-world Semantic Segmentation0
A Unified Framework for 3D Scene UnderstandingCode2
Understanding Multi-Granularity for Open-Vocabulary Part SegmentationCode2
Open-Vocabulary Semantic Segmentation with Image Embedding BalancingCode1
EAGLE: Efficient Adaptive Geometry-based Learning in Cross-view Understanding0
OpenDAS: Open-Vocabulary Domain Adaptation for 2D and 3D Segmentation0
Parameter-efficient Fine-tuning in Hyperspherical Space for Open-vocabulary Semantic Segmentation0
Pay Attention to Your Neighbours: Training-Free Open-Vocabulary Semantic SegmentationCode2
Training-Free Open-Vocabulary Segmentation with Offline Diffusion-Augmented Prototype Generation0
GOV-NeSF: Generalizable Open-Vocabulary Neural Semantic FieldsCode1
TTD: Text-Tag Self-Distillation Enhancing Image-Text Alignment in CLIP to Alleviate Single Tag BiasCode1
Image-to-Image Matching via Foundation Models: A New Perspective for Open-Vocabulary Semantic Segmentation0
MaskDiffusion: Exploiting Pre-trained Diffusion Models for Semantic SegmentationCode0
TAG: Guidance-free Open-Vocabulary Semantic SegmentationCode1
PosSAM: Panoptic Open-vocabulary Segment AnythingCode2
Multi-Grained Cross-modal Alignment for Learning Open-vocabulary Semantic Segmentation from Text Supervision0
Generalizable Semantic Vision Query Generation for Zero-shot Panoptic and Semantic Segmentation0
Bridging Generative and Discriminative Models for Unified Visual Perception with Diffusion Priors0
Exploring Simple Open-Vocabulary Semantic SegmentationCode1
UMG-CLIP: A Unified Multi-Granularity Vision Generalist for Open-World UnderstandingCode1
Open-Vocabulary 3D Semantic Segmentation with Foundation Models0
TagAlign: Improving Vision-Language Alignment with Multi-Tag ClassificationCode1
CLIP-DINOiser: Teaching CLIP a few DINO tricks for open-vocabulary semantic segmentationCode2
SkySense: A Multi-Modal Remote Sensing Foundation Model Towards Universal Interpretation for Earth Observation ImageryCode3
Open-Vocabulary Segmentation with Semantic-Assisted CalibrationCode1
Auto-Vocabulary Semantic SegmentationCode1
Emergent Open-Vocabulary Semantic Segmentation from Off-the-shelf Vision-Language ModelsCode1
SED: A Simple Encoder-Decoder for Open-Vocabulary Semantic SegmentationCode1
Uncovering Prototypical Knowledge for Weakly Open-Vocabulary Semantic SegmentationCode1
SILC: Improving Vision Language Pretraining with Self-Distillation0
OV-PARTS: Towards Open-Vocabulary Part SegmentationCode1
CLIPSelf: Vision Transformer Distills Itself for Open-Vocabulary Dense PredictionCode2
Learning Mask-aware CLIP Representations for Zero-Shot SegmentationCode1
CLIP-DIY: CLIP Dense Inference Yields Open-Vocabulary Semantic Segmentation For-FreeCode1
Panoptic Vision-Language Feature FieldsCode1
Diffusion Model is Secretly a Training-free Open Vocabulary Semantic SegmenterCode1
AttrSeg: Open-Vocabulary Semantic Segmentation via Attribute Decomposition-AggregationCode0
Convolutions Die Hard: Open-Vocabulary Segmentation with Single Frozen Convolutional CLIPCode2
Exploring Open-Vocabulary Semantic Segmentation without Human Labels0
SAD: Segment Any RGBDCode2
TagCLIP: Improving Discrimination Ability of Open-Vocabulary Semantic SegmentationCode1
MVP-SEG: Multi-View Prompt Learning for Open-Vocabulary Semantic Segmentation0
A Closer Look at the Explainability of Contrastive Language-Image Pre-trainingCode1
Open-Vocabulary Semantic Segmentation with Decoupled One-Pass NetworkCode1
CAT-Seg: Cost Aggregation for Open-Vocabulary Semantic SegmentationCode2
Global Knowledge Calibration for Fast Open-Vocabulary SegmentationCode1
Open-Vocabulary Panoptic Segmentation with Text-to-Image Diffusion ModelsCode2
Show:102550
← PrevPage 2 of 3Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1HyperSegmIoU64.6Unverified
2SILCmIoU63.5Unverified
3CAT-SegmIoU63.3Unverified
4MaskCLIP++mIoU62.5Unverified
5CLIPSelfmIoU62.3Unverified
6UMG-CLIP-L/14mIoU61Unverified
7SEDmIoU60.6Unverified
8Mask-AdaptermIoU60.4Unverified
9EBSeg-LmIoU60.2Unverified
10MAFT+mIoU59.4Unverified
#ModelMetricClaimedVerifiedStatus
1UMG-CLIP-E/14mIoU38.2Unverified
2MaskCLIP++mIoU38.2Unverified
3Mask-AdaptermIoU38.2Unverified
4CAT-SegmIoU37.9Unverified
5SILCmIoU37.7Unverified
6UMG-CLIP-L/14mIoU36.1Unverified
7MAFT+mIoU36.1Unverified
8OVSeg + OpenDASmIoU35.8Unverified
9SEDmIoU35.2Unverified
10CLIPSelfmIoU34.5Unverified
#ModelMetricClaimedVerifiedStatus
1UMG-CLIP-E/14mIoU17.3Unverified
2MaskCLIP++mIoU16.8Unverified
3Mask-AdaptermIoU16.2Unverified
4CAT-SegmIoU16Unverified
5UMG-CLIP-L/14mIoU15.4Unverified
6MAFT+mIoU15.1Unverified
7SILCmIoU15Unverified
8PosSAMmIoU14.9Unverified
9FC-CLIPmIoU14.8Unverified
10SCANmIoU14Unverified
#ModelMetricClaimedVerifiedStatus
1UMG-CLIP-L/14mIoU97.9Unverified
2SILCmIoU97.6Unverified
3SCANmIoU97.2Unverified
4CAT-SegmIoU97Unverified
5MaskCLIP++mIoU96.8Unverified
6MAFT+mIoU96.5Unverified
7EBSeg-LmIoU96.4Unverified
8FC-CLIPmIoU95.4Unverified
9OVSeg Swin-BmIoU94.5Unverified
10HyperSegmIoU92.1Unverified
#ModelMetricClaimedVerifiedStatus
1SILCmIoU25.8Unverified
2UMG-CLIP-E/14mIoU25.2Unverified
3MaskCLIP++mIoU23.9Unverified
4CAT-SegmIoU23.8Unverified
5UMG-CLIP-L/14mIoU23.2Unverified
6Mask-AdaptermIoU22.7Unverified
7SEDmIoU22.6Unverified
8MAFT+mIoU21.6Unverified
9EBSeg-LmIoU21Unverified
10FC-CLIPmIoU18.2Unverified
#ModelMetricClaimedVerifiedStatus
1POMPHIoU39.1Unverified
2ZSSegHIoU37.8Unverified
3ZegFormerHIoU34.8Unverified
4TTD (TCL)mIoU23.7Unverified
5LaVGmIoU23.2Unverified
6CLIP Surgery (original CLIP without any fine-tuning)mIoU21.9Unverified
7TTD (MaskCLIP)mIoU19.4Unverified
#ModelMetricClaimedVerifiedStatus
1FC-CLIPmIoU56.2Unverified
2SimSegmIoU34.5Unverified
3TTD (TCL)mIoU32Unverified
4CLIP Surgery (CLIP without any fine-tuning)mIoU31.4Unverified
5TTD (MaskCLIP)mIoU27Unverified
#ModelMetricClaimedVerifiedStatus
1UMG-CLIP-E/14mIoU85.4Unverified
2CAT-SegmIoU82.5Unverified
3SILCmIoU82.5Unverified
4FC-CLIPmIoU81.8Unverified
#ModelMetricClaimedVerifiedStatus
1SkySense-OmIoU-43.9Unverified
2SegEarth-OVmIoU-21.7Unverified
#ModelMetricClaimedVerifiedStatus
1PACLmIoU38.8Unverified
#ModelMetricClaimedVerifiedStatus
1SkySense-OmIoU8.3Unverified
#ModelMetricClaimedVerifiedStatus
1SkySense-OmIoU54.1Unverified
#ModelMetricClaimedVerifiedStatus
1SkySense-OmIoU30.89Unverified
#ModelMetricClaimedVerifiedStatus
1SkySense-OmIoU32.12Unverified