MSTA3D: Multi-scale Twin-attention for 3D Instance Segmentation
Duc Dang Trung Tran, Byeongkeun Kang, Yeejin Lee
Unverified — Be the first to reproduce this paper.
ReproduceAbstract
Recently, transformer-based techniques incorporating superpoints have become prevalent in 3D instance segmentation. However, they often encounter an over-segmentation problem, especially noticeable with large objects. Additionally, unreliable mask predictions stemming from superpoint mask prediction further compound this issue. To address these challenges, we propose a novel framework called MSTA3D. It leverages multi-scale feature representation and introduces a twin-attention mechanism to effectively capture them. Furthermore, MSTA3D integrates a box query with a box regularizer, offering a complementary spatial constraint alongside semantic queries. Experimental evaluations on ScanNetV2, ScanNet200 and S3DIS datasets demonstrate that our approach surpasses state-of-the-art 3D instance segmentation methods.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| S3DIS | MSTA3D | AP@50 | 70 | — | Unverified |
| ScanNet200 | MSTA3D | mAP | 26.2 | — | Unverified |
| ScanNetV2 | MSTA3D | mAP @ 50 | 79.5 | — | Unverified |