SOTAVerified

SAM2-UNet: Segment Anything 2 Makes Strong Encoder for Natural and Medical Image Segmentation

2024-08-16Code Available3· sign in to hype

Xinyu Xiong, Zihuang Wu, Shuangyi Tan, Wenxue Li, Feilong Tang, Ying Chen, Siying Li, Jie Ma, Guanbin Li

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Image segmentation plays an important role in vision understanding. Recently, the emerging vision foundation models continuously achieved superior performance on various tasks. Following such success, in this paper, we prove that the Segment Anything Model 2 (SAM2) can be a strong encoder for U-shaped segmentation models. We propose a simple but effective framework, termed SAM2-UNet, for versatile image segmentation. Specifically, SAM2-UNet adopts the Hiera backbone of SAM2 as the encoder, while the decoder uses the classic U-shaped design. Additionally, adapters are inserted into the encoder to allow parameter-efficient fine-tuning. Preliminary experiments on various downstream tasks, such as camouflaged object detection, salient object detection, marine animal segmentation, mirror detection, and polyp segmentation, demonstrate that our SAM2-UNet can simply beat existing specialized state-of-the-art methods without bells and whistles. Project page: https://github.com/WZH0120/SAM2-UNet.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
MAS3KSAM2-UNetmIoU0.8Unverified
MSD (Mirror Segmentation Dataset)SAM2-UNetIoU0.92Unverified
PMDSAM2-UNetIoU0.73Unverified
RMASSAM2-UNetmIoU0.74Unverified

Reproductions