Efficient Multi-Task RGB-D Scene Analysis for Indoor Environments
Daniel Seichter, Söhnke Benedikt Fischedick, Mona Köhler, Horst-Michael Groß
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/tui-nicr/emsanetOfficialIn paperpytorch★ 70
- github.com/tui-nicr/nicr-scene-analysis-datasetspytorch★ 17
Abstract
Semantic scene understanding is essential for mobile agents acting in various environments. Although semantic segmentation already provides a lot of information, details about individual objects as well as the general scene are missing but required for many real-world applications. However, solving multiple tasks separately is expensive and cannot be accomplished in real time given limited computing and battery capabilities on a mobile platform. In this paper, we propose an efficient multi-task approach for RGB-D scene analysis~(EMSANet) that simultaneously performs semantic and instance segmentation~(panoptic segmentation), instance orientation estimation, and scene classification. We show that all tasks can be accomplished using a single neural network in real time on a mobile platform without diminishing performance - by contrast, the individual tasks are able to benefit from each other. In order to evaluate our multi-task approach, we extend the annotations of the common RGB-D indoor datasets NYUv2 and SUNRGB-D for instance segmentation and orientation estimation. To the best of our knowledge, we are the first to provide results in such a comprehensive multi-task setting for indoor scene analysis on NYUv2 and SUNRGB-D.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| NYU-Depth V2 | EMSANet | PQ | 47.38 | — | Unverified |
| SUN-RGBD | EMSANet | PQ | 52.84 | — | Unverified |