SOTAVerified

Semantic Segmentation

Papers

Showing 626650 of 14763 papers

TitleStatusHype
A Unified Framework for 3D Scene UnderstandingCode2
DatasetDM: Synthesizing Data with Perception Annotations Using Diffusion ModelsCode2
In Defense of Lazy Visual Grounding for Open-Vocabulary Semantic SegmentationCode2
AllWeatherNet:Unified Image Enhancement for Autonomous Driving under Adverse Weather and Lowlight-conditionsCode2
A Unified Transformer Framework for Group-based Segmentation: Co-Segmentation, Co-Saliency Detection and Video Salient Object DetectionCode2
Fast-iTPN: Integrally Pre-Trained Transformer Pyramid Network with Token MigrationCode2
Customized Segment Anything Model for Medical Image SegmentationCode2
Dataset QuantizationCode2
InvPT: Inverted Pyramid Multi-task Transformer for Dense Scene UnderstandingCode2
KPConvX: Modernizing Kernel Point Convolution with Kernel AttentionCode2
Label Anything: Multi-Class Few-Shot Semantic Segmentation with Visual PromptsCode2
LambdaNetworks: Modeling Long-Range Interactions Without AttentionCode2
DeepGCNs: Making GCNs Go as Deep as CNNsCode2
Alleviating Textual Reliance in Medical Language-guided Segmentation via Prototype-driven Semantic ApproximationCode2
Boundary-Aware Segmentation Network for Mobile and Web ApplicationsCode2
Learning Affinity from Attention: End-to-End Weakly-Supervised Semantic Segmentation with TransformersCode2
Learning Embeddings with Centroid Triplet Loss for Object Identification in Robotic GraspingCode2
Learning Semantic Segmentation of Large-Scale Point Clouds with Random SamplingCode2
Learning Vision from Models Rivals Learning Vision from DataCode2
Learning without Exact Guidance: Updating Large-scale High-resolution Land Cover Maps from Low-resolution Historical LabelsCode2
Learn to Rectify the Bias of CLIP for Unsupervised Semantic SegmentationCode2
Cross-Image Relational Knowledge Distillation for Semantic SegmentationCode2
CrossEarth: Geospatial Vision Foundation Model for Domain Generalizable Remote Sensing Semantic SegmentationCode2
Locality Alignment Improves Vision-Language ModelsCode2
CrossFormer++: A Versatile Vision Transformer Hinging on Cross-scale AttentionCode2
Show:102550
← PrevPage 26 of 591Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1InternImage-H (M3I Pre-training)Params (M)1,310Unverified
2ViT-P (InternImage-H)Validation mIoU63.6Unverified
3ONE-PEACEValidation mIoU63Unverified
4M3I Pre-training (InternImage-H)Validation mIoU62.9Unverified
5InternImage-HValidation mIoU62.9Unverified
6BEiT-3Validation mIoU62.8Unverified
7EVAValidation mIoU62.3Unverified
8ViT-P (OneFormer, InternImage-H)Validation mIoU61.6Unverified
9ViT-Adapter-L (Mask2Former, BEiTv2 pretrain)Validation mIoU61.5Unverified
10FD-SwinV2-GValidation mIoU61.4Unverified