NeRFLiX: High-Quality Neural View Synthesis by Learning a Degradation-Driven Inter-viewpoint MiXer
Kun Zhou, Wenbo Li, Yi Wang, Tao Hu, Nianjuan Jiang, Xiaoguang Han, Jiangbo Lu
Code Available — Be the first to reproduce this paper.
ReproduceCode
- github.com/redrock303/NeRFLiX_CPVR2023Officialpytorch★ 107
Abstract
Neural radiance fields (NeRF) show great success in novel view synthesis. However, in real-world scenes, recovering high-quality details from the source images is still challenging for the existing NeRF-based approaches, due to the potential imperfect calibration information and scene representation inaccuracy. Even with high-quality training frames, the synthetic novel views produced by NeRF models still suffer from notable rendering artifacts, such as noise, blur, etc. Towards to improve the synthesis quality of NeRF-based approaches, we propose NeRFLiX, a general NeRF-agnostic restorer paradigm by learning a degradation-driven inter-viewpoint mixer. Specially, we design a NeRF-style degradation modeling approach and construct large-scale training data, enabling the possibility of effectively removing NeRF-native rendering artifacts for existing deep neural networks. Moreover, beyond the degradation removal, we propose an inter-viewpoint aggregation framework that is able to fuse highly related high-quality training images, pushing the performance of cutting-edge NeRF models to entirely new levels and producing highly photo-realistic synthetic views.
Tasks
Benchmark Results
| Dataset | Model | Metric | Claimed | Verified | Status |
|---|---|---|---|---|---|
| LLFF | TensoRF + NeRFLiX | PSNR | 27.39 | — | Unverified |
| LLFF | Plenoxels + NeRFLiX | PSNR | 26.9 | — | Unverified |
| Tanks and Temples | TensoRF + NeRFLiX | PSNR | 28.94 | — | Unverified |
| Tanks and Temples | Plenoxels + NeRFLiX | PSNR | 28.61 | — | Unverified |
| Tanks and Temples | DIVeR + NeRFLiX | SSIM | 0.92 | — | Unverified |