Large-scale Point Cloud Semantic Segmentation with Superpoint Graphs
Loic Landrieu, Martin Simonovsky
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/loicland/superpoint_graphOfficialIn paperpytorch★ 0
- github.com/jsgaobiao/superpoint_graphpytorch★ 0
Abstract
We propose a novel deep learning-based framework to tackle the challenge of semantic segmentation of large-scale point clouds of millions of points. We argue that the organization of 3D point clouds can be efficiently captured by a structure called superpoint graph (SPG), derived from a partition of the scanned scene into geometrically homogeneous elements. SPGs offer a compact yet rich representation of contextual relationships between object parts, which is then exploited by a graph convolutional network. Our framework sets a new state of the art for segmenting outdoor LiDAR scans (+11.9 and +8.8 mIoU points for both Semantic3D test sets), as well as indoor scans (+12.4 mIoU points for the S3DIS dataset).
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| DALES | SPG | mIoU | 60.6 | — | Unverified |
| SemanticKITTI | SPGraph | test mIoU | 17.4 | — | Unverified |
| SensatUrban | SPGraph | mIoU | 37.29 | — | Unverified |