SOTAVerified

Depth-Adapted CNNs for RGB-D Semantic Segmentation

2022-06-08Unverified0· sign in to hype

Zongwei Wu, Guillaume Allibert, Christophe Stolz, Chao Ma, Cédric Demonceaux

Unverified — Be the first to reproduce this paper.

Reproduce

Abstract

Recent RGB-D semantic segmentation has motivated research interest thanks to the accessibility of complementary modalities from the input side. Existing works often adopt a two-stream architecture that processes photometric and geometric information in parallel, with few methods explicitly leveraging the contribution of depth cues to adjust the sampling position on RGB images. In this paper, we propose a novel framework to incorporate the depth information in the RGB convolutional neural network (CNN), termed Z-ACN (Depth-Adapted CNN). Specifically, our Z-ACN generates a 2D depth-adapted offset which is fully constrained by low-level features to guide the feature extraction on RGB images. With the generated offset, we introduce two intuitive and effective operations to replace basic CNN operators: depth-adapted convolution and depth-adapted average pooling. Extensive experiments on both indoor and outdoor semantic segmentation tasks demonstrate the effectiveness of our approach.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
NYU-Depth V2Z-ACN (ResNet-101)Mean IoU51.24Unverified
NYU-Depth V2Z-ACN (ResNet-50)Mean IoU50.05Unverified
NYU-Depth V2Z-ACN (ResNet-34)Mean IoU49.15Unverified
NYU-Depth V2Z-ACN (ResNet-18)Mean IoU47.02Unverified

Reproductions