SOTAVerified

Inverse Rendering

Inverse Rendering is the task of recovering the properties of a scene, such as shape, material, and lighting, from an image or a video. The goal of inverse rendering is to determine the properties of a scene given an observation of it, and to generate new images or videos based on these properties.

Papers

Showing 201210 of 271 papers

TitleStatusHype
Deep Face Feature for Face Alignment0
Deep Generative Models: Deterministic Prediction with an Application in Inverse Rendering0
Deep Learning compatible Differentiable X-ray Projections for Inverse Rendering0
Deep Polarization Cues for Single-shot Shape and Subsurface Scattering Estimation0
DeepShaRM: Multi-View Shape and Reflectance Map Recovery Under Unknown Lighting0
Deep Structure for end-to-end inverse rendering0
Deep Uncalibrated Photometric Stereo via Inter-Intra Image Feature Fusion0
DEL: Discrete Element Learner for Learning 3D Particle Dynamics with Neural Rendering0
DiffCSG: Differentiable CSG via Rasterization0
Differentiable Inverse Rendering with Interpretable Basis BRDFs0
Show:102550
← PrevPage 21 of 28Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Neural-PBIRHDR-PSNR26.01Unverified
2NVDiffRecMCHDR-PSNR24.43Unverified
3InvRenderHDR-PSNR23.76Unverified
4NeRFactorHDR-PSNR23.54Unverified
5NeRDHDR-PSNR23.29Unverified
6NVDiffRecHDR-PSNR22.91Unverified
7PhySGHDR-PSNR21.81Unverified