PU-Net: Point Cloud Upsampling Network
Lequan Yu, Xianzhi Li, Chi-Wing Fu, Daniel Cohen-Or, Pheng-Ann Heng
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/yulequan/PU-NetOfficialIn papertf★ 0
- github.com/guochengqian/PU-GCNtf★ 0
- github.com/skoo9500/3d-pc-AE-GANtf★ 0
Abstract
Learning and analyzing 3D point clouds with deep networks is challenging due to the sparseness and irregularity of the data. In this paper, we present a data-driven point cloud upsampling technique. The key idea is to learn multi-level features per point and expand the point set via a multi-branch convolution unit implicitly in feature space. The expanded feature is then split to a multitude of features, which are then reconstructed to an upsampled point set. Our network is applied at a patch-level, with a joint loss function that encourages the upsampled points to remain on the underlying surface with a uniform distribution. We conduct various experiments using synthesis and scan data to evaluate our method and demonstrate its superiority over some baseline methods and an optimization-based method. Results show that our upsampled points have better uniformity and are located closer to the underlying surfaces.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| SHREC15 | PU-NET | F-measure (%) | 56.4 | — | Unverified |