SOTAVerified

Depth-aware CNN for RGB-D Segmentation

2018-03-19ECCV 2018Code Available0· sign in to hype

Weiyue Wang, Ulrich Neumann

Code Available — Be the first to reproduce this paper.

Reproduce

Code

Abstract

Convolutional neural networks (CNN) are limited by the lack of capability to handle geometric information due to the fixed grid kernel structure. The availability of depth data enables progress in RGB-D semantic segmentation with CNNs. State-of-the-art methods either use depth as additional images or process spatial information in 3D volumes or point clouds. These methods suffer from high computation and memory cost. To address these issues, we present Depth-aware CNN by introducing two intuitive, flexible and effective operations: depth-aware convolution and depth-aware average pooling. By leveraging depth similarity between pixels in the process of information propagation, geometry is seamlessly incorporated into CNN. Without introducing any additional parameters, both operators can be easily integrated into existing CNNs. Extensive experiments and ablation studies on challenging RGB-D semantic segmentation benchmarks validate the effectiveness and flexibility of our approach.

Tasks

Benchmark Results

DatasetModelMetricClaimedVerifiedStatus
NYU-Depth V2Depth-aware CNNMean IoU43.9Unverified
Stanford2D3D - RGBDDepth-aware CNNmIoU39.5Unverified
SUN-RGBDTokenFusion (S)Mean IoU45.73Unverified
SUN-RGBDTokenFusion (S)Mean IoU50Unverified
SUN-RGBDTokenFusion (S)Mean IoU53Unverified
SUN-RGBDTokenFusion (S)Mean IoU42Unverified
SUN-RGBDTokenFusion (S)Mean IoU48.6Unverified

Reproductions