SOTAVerified

Inverse Rendering

Inverse Rendering is the task of recovering the properties of a scene, such as shape, material, and lighting, from an image or a video. The goal of inverse rendering is to determine the properties of a scene given an observation of it, and to generate new images or videos based on these properties.

Papers

Showing 211220 of 271 papers

TitleStatusHype
IRISformer: Dense Vision Transformers for Single-Image Inverse Rendering in Indoor ScenesCode0
TileGen: Tileable, Controllable Material Generation and Capture0
Differentiable Rendering of Neural SDFs through Reparameterization0
Rotation-Equivariant Conditional Spherical Neural Fields for Learning a Natural Illumination Prior0
Physically-Based Editing of Indoor Scene Lighting from a Single Image0
Physically Disentangled RepresentationsCode0
IRON: Inverse Rendering by Optimizing Neural SDFs and Materials from Photometric Images0
Dressi: A Hardware-Agnostic Differentiable Renderer with Reactive Shader Packing and Soft Rasterization0
FIRe: Fast Inverse Rendering using Directional and Signed Distance Functions0
PANDORA: Polarization-Aided Neural Decomposition Of Radiance0
Show:102550
← PrevPage 22 of 28Next →

Benchmark Results

#ModelMetricClaimedVerifiedStatus
1Neural-PBIRHDR-PSNR26.01Unverified
2NVDiffRecMCHDR-PSNR24.43Unverified
3InvRenderHDR-PSNR23.76Unverified
4NeRFactorHDR-PSNR23.54Unverified
5NeRDHDR-PSNR23.29Unverified
6NVDiffRecHDR-PSNR22.91Unverified
7PhySGHDR-PSNR21.81Unverified