Depth Map Prediction from a Single Image using a Multi-Scale Deep Network
David Eigen, Christian Puhrsch, Rob Fergus
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/SeokjuLee/Insta-DMpytorch★ 228
- github.com/MasazI/cnn_depth_tensorflowtf★ 127
- github.com/sejong-rcv/2021.Paper.TransDSSLpytorch★ 12
- github.com/kieran514/dyna-dmpytorch★ 12
- github.com/zhengtr/CS231N-Final-Projectpytorch★ 0
- github.com/dsshim0125/daclpytorch★ 0
- github.com/imran3180/depth-map-predictionpytorch★ 0
- github.com/Tom-Zheng/depth_single_imagetf★ 0
- github.com/mindspore-ai/models/tree/master/official/cv/depthnetmindspore★ 0
- github.com/shuuchen/depth_eigenpytorch★ 0
Abstract
Predicting depth is an essential component in understanding the 3D geometry of a scene. While for stereo images local correspondence suffices for estimation, finding depth relations from a single image is less straightforward, requiring integration of both global and local information from various cues. Moreover, the task is inherently ambiguous, with a large source of uncertainty coming from the overall scale. In this paper, we present a new method that addresses this task by employing two deep network stacks: one that makes a coarse global prediction based on the entire image, and another that refines this prediction locally. We also apply a scale-invariant error to help measure depth relations rather than scale. By leveraging the raw datasets as large sources of training data, our method achieves state-of-the-art results on both NYU Depth and KITTI, and matches detailed depth boundaries without the need for superpixelation.