SOTAVerified

The Perfect Match: 3D Point Cloud Matching with Smoothed Densities

2018-11-16CVPR 2019Code Available0· sign in to hype

Zan Gojcic, Caifa Zhou, Jan D. Wegner, Andreas Wieser

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

We propose 3DSmoothNet, a full workflow to match 3D point clouds with a siamese deep learning architecture and fully convolutional layers using a voxelized smoothed density value (SDV) representation. The latter is computed per interest point and aligned to the local reference frame (LRF) to achieve rotation invariance. Our compact, learned, rotation invariant 3D point cloud descriptor achieves 94.9% average recall on the 3DMatch benchmark data set, outperforming the state-of-the-art by more than 20 percent points with only 32 output dimensions. This very low output dimension allows for near realtime correspondence search with 0.1 ms per feature point on a standard PC. Our approach is sensor- and sceneagnostic because of SDV, LRF and learning highly descriptive features with fully convolutional layers. We show that 3DSmoothNet trained only on RGB-D indoor scenes of buildings achieves 79.0% average recall on laser scans of outdoor vegetation, more than double the performance of our closest, learning-based competitors. Code, data and pre-trained models are available online at https://github.com/zgojcic/3DSmoothNet.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
3DLoMatch (10-30% overlap)3DSN (reported in PREDATOR)Recall ( correspondence RMSE below 0.2)33Unverified
3DMatch (at least 30% overlapped - sample 5k interest points)3DSN (reported in PREDATOR)Recall ( correspondence RMSE below 0.2)78.4Unverified
3DMatch Benchmark3DSmoothNetFeature Matching Recall94.7Unverified
ETH (trained on 3DMatch)PerfectMatchFeature Matching Recall0.79Unverified

Reproductions