SOTAVerified

Depth Estimation

Depth Estimation is the task of measuring the distance of each pixel relative to the camera. Depth is extracted from either monocular (single) or stereo (multiple views of a scene) images. Traditional methods use multi-view geometry to find the relationship between the images. Newer methods can directly estimate depth by minimizing the regression loss, or by learning to generate a novel view from a sequence. The most popular benchmarks are KITTI and NYUv2. Models are typically evaluated according to a RMS metric.

Source: DIODE: A Dense Indoor and Outdoor DEpth Dataset

Papers

Showing 11011150 of 2454 papers

TitleStatusHype
Energy-Efficient Adaptive 3D SensingCode1
NeuralLift-360: Lifting an In-the-Wild 2D Photo to a 3D Object With 360deg Views0
OmniVidar: Omnidirectional Depth Estimation From Multi-Fisheye ImagesCode1
Depth Estimation From Camera Image and mmWave Radar Point Cloud0
AttEntropy: On the Generalization Ability of Supervised Semantic Segmentation Transformers to New Objects in New DomainsCode0
HandsOff: Labeled Dataset Generation With No Additional Human Annotations0
Shakes on a Plane: Unsupervised Depth Estimation from Unstabilized Photography0
Vision-Based Environmental Perception for Autonomous Driving0
Depth Estimation maps of lidar and stereo images0
MaskingDepth: Masked Consistency Regularization for Semi-supervised Monocular Depth EstimationCode1
Lightweight Monocular Depth Estimation0
Scene-aware Egocentric 3D Human Pose EstimationCode1
DAG: Depth-Aware Guidance with Denoising Diffusion Probabilistic ModelsCode1
Are We Ready for Vision-Centric Driving Streaming Perception? The ASAP BenchmarkCode1
Solve the Puzzle of Instance Segmentation in Videos: A Weakly Supervised Framework with Spatio-Temporal Collaboration0
Towards Practical Plug-and-Play Diffusion ModelsCode1
ROIFormer: Semantic-Aware Region of Interest Transformer for Efficient Self-Supervised Monocular Depth Estimation0
Mind The Edge: Refining Depth Edges in Sparsely-Supervised Monocular Depth EstimationCode1
Cross-Domain Synthetic-to-Real In-the-Wild Depth and Normal Estimation for 3D Scene Understanding0
Objects as Spatio-Temporal 2.5D points0
Event-based Monocular Dense Depth Estimation with Recurrent Transformers0
GARF:Geometry-Aware Generalized Neural Radiance Field0
SceneRF: Self-Supervised Monocular 3D Scene Reconstruction with Radiance FieldsCode2
3D Object Aided Self-Supervised Monocular Depth Estimation0
Multi-resolution Monocular Depth Map Fusion by Self-supervised Gradient-based CompositionCode1
Geometry-Aware Network for Domain Adaptive Semantic Segmentation0
BEV-LGKD: A Unified LiDAR-Guided Knowledge Distillation Framework for BEV 3D Object DetectionCode1
Weakly Supervised 3D Multi-person Pose Estimation for Large-scale Scenes based on Monocular Camera and Single LiDAR0
BEVUDA: Multi-geometric Space Alignments for Domain Adaptive BEV 3D Object Detection0
ObjCAViT: Improving Monocular Depth Estimation Using Natural Language Models And Image-Object Cross-AttentionCode0
Attention-Based Depth Distillation with 3D-Aware Positional Encoding for Monocular 3D Object DetectionCode0
Self-Supervised Surround-View Depth Estimation with Volumetric Feature FusionCode1
Generalized Face Anti-Spoofing via Multi-Task Learning and One-Side Meta Triplet Loss0
SuperFusion: Multilevel LiDAR-Camera Fusion for Long-Range HD Map GenerationCode2
3DPPE: 3D Point Positional Encoding for Multi-Camera 3D Object Detection TransformersCode1
Copy-Pasting Coherent Depth Regions Improves Contrastive Learning for Urban-Scene SegmentationCode0
How do Cross-View and Cross-Modal Alignment Affect Representations in Contrastive Learning?0
Lite-Mono: A Lightweight CNN and Transformer Architecture for Self-Supervised Monocular Depth EstimationCode2
The Monocular Depth Estimation ChallengeCode1
Event Transformer+. A multi-purpose solution for efficient event data processing0
AeDet: Azimuth-invariant Multi-view 3D Object DetectionCode1
MSS-DepthNet: Depth Prediction with Multi-Step Spiking Neural Network0
Dynamic Depth-Supervised NeRF for Multi-View RGB-D Operating Room Images0
DesNet: Decomposed Scale-Consistent Network for Unsupervised Depth Completion0
Hybrid Transformer Based Feature Fusion for Self-Supervised Monocular Depth Estimation0
A Practical Stereo Depth System for Smart GlassesCode1
Improving Pixel-Level Contrastive Learning by Leveraging Exogenous Depth Information0
BEVDistill: Cross-Modal BEV Distillation for Multi-View 3D Object DetectionCode1
LightDepth: A Resource Efficient Depth Estimation Approach for Dealing with Ground Truth Sparsity via Curriculum LearningCode1
Reconfigurable Intelligent Surface Aided Wireless Sensing for Scene Depth EstimationCode0
Show:102550
← PrevPage 23 of 50Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1OmniDepthRMSE0.62Unverified
2SphereDepthRMSE0.45Unverified
3Jin et al.RMSE0.42Unverified
4BiFuse with fusionRMSE0.41Unverified
5HoHoNet (ResNet-101)RMSE0.38Unverified
6PanoDepthRMSE0.37Unverified
7BiFuse++RMSE0.37Unverified
8UniFuse with fusionRMSE0.37Unverified
9DisConvRMSE0.37Unverified
10SliceNetRMSE0.37Unverified
#ModelMetricClaimedVerifiedStatus
1A2JmAP8.61Unverified
2PAD-NetRMS0.79Unverified
3MS-CRFRMS0.59Unverified
4DORNRMS0.51Unverified
5FreeformRMS0.43Unverified
6Optimized, freeformRMS0.43Unverified
7VNLRMS0.42Unverified
8BTSRMS0.41Unverified
9TransDepth (AGD+ ViT)RMS0.37Unverified
10AdaBinsRMS0.36Unverified
#ModelMetricClaimedVerifiedStatus
1T2NetAbs Rel0.35Unverified
2MIDASAbs Rel0.31Unverified
3Bhattacharjee et al.Abs Rel0.25Unverified
#ModelMetricClaimedVerifiedStatus
1T2NetAbs Rel0.49Unverified
2MIDASAbs Rel0.42Unverified
3Bhattacharjee et al.Abs Rel0.38Unverified
#ModelMetricClaimedVerifiedStatus
1LeReSabsolute relative error0.1Unverified
2DELTASabsolute relative error0.09Unverified
3Distill Any Depthabsolute relative error0.04Unverified
#ModelMetricClaimedVerifiedStatus
1SDC-DepthRMSE6.92Unverified
2SwinMTLRMSE6.35Unverified
#ModelMetricClaimedVerifiedStatus
1AIP-BrownDelta < 1.250.36Unverified
2LeResDelta < 1.250.23Unverified
#ModelMetricClaimedVerifiedStatus
1H-Net (Ours)Absolute relative error (AbsRel)0.09Unverified
2H-Net (Ours) Full EigenAbsolute relative error (AbsRel)0.08Unverified
#ModelMetricClaimedVerifiedStatus
1GLPDepthDelta < 1.250.43Unverified
2SRDINET (Model A)Delta < 1.250.4Unverified
#ModelMetricClaimedVerifiedStatus
1Atlas (finetuned)RMSE0.17Unverified
2Atlas (plain)RMSE0.17Unverified
#ModelMetricClaimedVerifiedStatus
1LFattNetBadPix(0.01)17.23Unverified
#ModelMetricClaimedVerifiedStatus
1LightDepthNumber of parameters (M)42.6Unverified
#ModelMetricClaimedVerifiedStatus
1UniFuseAbs Rel0.11Unverified
#ModelMetricClaimedVerifiedStatus
1X-TC (Cross-Task Consistency)L1 error1.63Unverified