SOTAVerified

MaskCLIP++: A Mask-Based CLIP Fine-tuning Framework for Open-Vocabulary Image Segmentation

2024-12-16Code Available1· sign in to hype

Quan-Sheng Zeng, Yunheng Li, Daquan Zhou, Guanbin Li, Qibin Hou, Ming-Ming Cheng

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Open-vocabulary image segmentation has been advanced through the synergy between mask generators and vision-language models like Contrastive Language-Image Pre-training (CLIP). Previous approaches focus on generating masks while aligning mask features with text embeddings during training. In this paper, we observe that relying on generated low-quality masks can weaken the alignment of vision and language in regional representations. This motivates us to present a new fine-tuning framework, named MaskCLIP++, which uses ground-truth masks instead of generated masks to enhance the mask classification capability of CLIP. Due to the limited diversity of image segmentation datasets with mask annotations, we propose incorporating a consistency alignment constraint during fine-tuning, which alleviates categorical bias toward the fine-tuning dataset. After low-cost fine-tuning, combining with the mask generator in previous state-of-the-art mask-based open vocabulary segmentation methods, we achieve performance improvements of +1.7, +2.3, +2.1, +3.1, and +0.3 mIoU on the A-847, PC-459, A-150, PC-59, and PAS-20 datasets, respectively. Code is released at https://github.com/HVision-NKU/MaskCLIPpp .

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
ADE20K-150MaskCLIP++mIoU38.2Unverified
ADE20K-847MaskCLIP++mIoU16.8Unverified
PASCAL Context-459MaskCLIP++mIoU23.9Unverified
PASCAL Context-59MaskCLIP++mIoU62.5Unverified
PascalVOC-20MaskCLIP++mIoU96.8Unverified

Reproductions