SOTAVerified

Monocular Depth Estimation

Monocular Depth Estimation is the task of estimating the depth value (distance relative to the camera) of each pixel given a single (monocular) RGB image. This challenging task is a key prerequisite for determining scene understanding for applications such as 3D scene reconstruction, autonomous driving, and AR. State-of-the-art methods usually fall into one of two categories: designing a complex network that is powerful enough to directly regress the depth map, or splitting the input into bins or windows to reduce computational complexity. The most popular benchmarks are the KITTI and NYUv2 datasets. Models are typically evaluated using RMSE or absolute relative error.

Source: Defocus Deblurring Using Dual-Pixel Data

Papers

Showing 110 of 876 papers

TitleStatusHype
Depth Pro: Sharp Monocular Metric Depth in Less Than a SecondCode9
Depth Anything V2Code9
Depth Anything: Unleashing the Power of Large-Scale Unlabeled DataCode9
Marigold: Affordable Adaptation of Diffusion-Based Image Generators for Image AnalysisCode7
Endo-4DGS: Endoscopic Monocular Scene Reconstruction with 4D Gaussian SplattingCode7
DINOv2: Learning Robust Visual Features without SupervisionCode6
UniDepthV2: Universal Monocular Metric Depth Estimation Made SimplerCode5
Video Depth Anything: Consistent Depth Estimation for Super-Long VideosCode5
DepthCrafter: Generating Consistent Long Depth Sequences for Open-world VideosCode5
UniDepth: Universal Monocular Metric Depth EstimationCode5
Show:102550
← PrevPage 1 of 88Next →

No leaderboard results yet.