Pano-NeRF: Synthesizing High Dynamic Range Novel Views with Geometry from Sparse Low Dynamic Range Panoramic Images (2312.15942v2)
Abstract: Panoramic imaging research on geometry recovery and High Dynamic Range (HDR) reconstruction becomes a trend with the development of Extended Reality (XR). Neural Radiance Fields (NeRF) provide a promising scene representation for both tasks without requiring extensive prior data. However, in the case of inputting sparse Low Dynamic Range (LDR) panoramic images, NeRF often degrades with under-constrained geometry and is unable to reconstruct HDR radiance from LDR inputs. We observe that the radiance from each pixel in panoramic images can be modeled as both a signal to convey scene lighting information and a light source to illuminate other pixels. Hence, we propose the irradiance fields from sparse LDR panoramic images, which increases the observation counts for faithful geometry recovery and leverages the irradiance-radiance attenuation for HDR reconstruction. Extensive experiments demonstrate that the irradiance fields outperform state-of-the-art methods on both geometry recovery and HDR reconstruction and validate their effectiveness. Furthermore, we show a promising byproduct of spatially-varying lighting estimation. The code is available at https://github.com/Lu-Zhan/Pano-NeRF.
- HRDFuse: Monocular 360deg depth estimation by collaboratively learning holistic-with-regional depth distributions. In CVPR, 13273–13282.
- Arrighetti, W. 2017. The academy color encoding system (ACES): A professional color-management framework for production, post-production and archival of still and motion pictures. Journal of Imaging, 3(4): 40.
- Azimi, M.; et al. 2021. PU21: A novel perceptually uniform encoding for adapting existing quality metrics for HDR. In 2021 Picture Coding Symposium (PCS), 1–5. IEEE.
- Mip-NeRF: A multiscale representation for anti-aliasing neural radiance fields. In ICCV, 5855–5864.
- Mip-NeRF 360: Unbounded anti-aliased neural radiance fields. In CVPR, 5470–5479.
- Neural reflectance fields for appearance acquisition. arXiv preprint arXiv:2008.03824.
- Nerd: Neural reflectance decomposition from image collections. In ICCV, 12684–12694.
- Depth estimation from indoor panoramas with neural scene representation. In CVPR, 899–908.
- MVSNeRF: Fast generalizable radiance field reconstruction from multi-view stereo. In ICCV, 14124–14133.
- HDR video reconstruction: A coarse-to-fine network and a real-world benchmark dataset. In ICCV, 2502–2511.
- Text2Light: Zero-shot text-driven HDR panorama generation. ACM TOG, 41(6).
- Stereo radiance fields (SRF): Learning view synthesis for sparse views of novel scenes. In CVPR, 7911–7920.
- 3D scene geometry estimation from 360 imagery: A survey. ACM Computing Surveys, 55(4): 1–39.
- Modeling and rendering architecture from photographs: A hybrid geometry-and image-based approach. In SIGGRAPH, 11–20.
- Depth-supervised NeRF: Fewer views and faster training for free. In CVPR, 12882–12891.
- Casual indoor HDR radiance capture from omnidirectional images. In BMVC.
- HDR-NeRF: High dynamic range neural radiance fields. In CVPR, 18398–18408.
- All-around depth from small motion with a spherical panoramic camera. In ECCV, 156–172. Springer.
- Putting NeRF on a diet: Semantically consistent few-shot view synthesis. In ICCV, 5885–5894.
- HDR-Plenoxels: Self-calibrating high dynamic range radiance fields. In ECCV, 384–401. Springer.
- Kajiya, J. T. 1986. The rendering equation. In SIGGRAPH, 143–150.
- Ray tracing volume densities. ACM SIGGRAPH computer graphics, 18(3): 165–174.
- Adam: A method for stochastic optimization. In ICLR.
- 360FusionNeRF: Panoramic neural radiance fields with joint guidance. In IROS, 7202–7209.
- HDRNET: Single-image-based HDR reconstruction using channel attention cnn. In ICMSSP, 119–124.
- Lighting, reflectance and geometry estimation from 360 panoramic stereo. In CVPR, 10586–10595.
- Omnifusion: 360 monocular depth estimation via geometry-aware fusion. In CVPR, 2801–2810.
- Inverse rendering for complex indoor scenes: Shape, spatially-varying lighting and svbrdf from a single image. In CVPR, 2475–2484.
- Multi-view inverse rendering for large-scale real-world indoor scenes. In CVPR, 12499–12509.
- Single-image HDR reconstruction by learning to reverse the camera pipeline. In CVPR, 1651–1660.
- HDR-VDP-3: A multi-metric for predicting image differences, quality and contrast distortions in high dynamic range and regular content. arXiv preprint arXiv:2304.13625.
- NeRF in the dark: High dynamic range view synthesis from noisy raw images. In CVPR, 16190–16199.
- NeRF: Representing scenes as neural radiance fields for view synthesis. Communications of the ACM, 65(1): 99–106.
- RegNeRF: Regularizing neural radiance fields for view synthesis from sparse inputs. In CVPR, 5480–5490.
- Ldr2Hdr: on-the-fly reverse tone mapping of legacy video and photographs. ACM TOG, 26(3).
- Dense depth priors for neural radiance fields from sparse input views. In CVPR, 12892–12901.
- Single image HDR reconstruction using a CNN with masked features and perceptual loss. ACM TOG, 39(4): 80–1.
- Generalized random walks for fusion of multi-exposure images. IEEE TIP, 20(12): 3634–3646.
- ViP-NeRF: Visibility prior for sparse input neural radiance fields. In ACM SIGGRAPH.
- NeRV: Neural reflectance and visibility fields for relighting and view synthesis. In CVPR, 7495–7504.
- Lighthouse: Predicting lighting volumes for spatially-coherent illumination. In CVPR, 8080–8089.
- The Replica dataset: A digital replica of indoor spaces. arXiv preprint arXiv:1906.05797.
- Volume rendering digest (for NeRF). arXiv preprint arXiv:2209.02417.
- Ref-NeRF: Structured view-dependent appearance for neural radiance fields. In CVPR, 5491–5500.
- Bifuse: Monocular 360 depth estimation via bi-projection fusion. In CVPR, 462–471.
- 360SD-Net: 360 stereo depth estimation with learnable cost volume. In 2020 IEEE ICRA, 582–588.
- Image quality assessment: from error visibility to structural similarity. IEEE TIP, 13(4): 600–612.
- Learning indoor inverse rendering with 3D spatially-varying lighting. In ICCV, 12538–12547.
- Rendering with radiance: The art and science of lighting visualization.
- DiffusioNeRF: Regularizing neural radiance fields with denoising diffusion models. In CVPR.
- Automatic spatially varying illumination recovery of indoor scenes based on a single RGB-D image. IEEE TVCG, 26(4): 1672–1685.
- FreeNeRF: Improving few-shot neural rendering with free frequency regularization. In CVPR, 8254–8263.
- NeILF: Neural incident light field for physically-based material estimation. In ECCV, 700–716. Springer.
- pixelNeRF: Neural radiance fields from one or few images. In CVPR, 4578–4587.
- Luminance attentive networks for HDR image and panorama reconstruction. In Computer Graphics Forum, volume 40, 181–192. Wiley Online Library.
- Improving 360 monocular depth estimation via non-local dense prediction transformer and joint supervised and self-supervised learning. In AAAI, volume 36, 3224–3233.
- Unsupervised depth estimation from monocular videos with hybrid geometric-refined loss and contextual attention. Neurocomputing, 379: 250–261.
- The unreasonable effectiveness of deep features as a perceptual metric. In CVPR.
- NeRFactor: Neural factorization of shape and reflectance under an unknown illumination. ACM TOG, 40(6): 1–18.
- Modeling indirect illumination for inverse rendering. In CVPR, 18643–18652.
- Glosh: Global-local spherical harmonics for intrinsic image decomposition. In ICCV, 7820–7829.
- ACDNet: Adaptively combined dilated convolution for monocular panorama depth estimation. In AAAI, volume 36, 3653–3661.