Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
139 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

BAD-NeRF: Bundle Adjusted Deblur Neural Radiance Fields (2211.12853v2)

Published 23 Nov 2022 in cs.CV

Abstract: Neural Radiance Fields (NeRF) have received considerable attention recently, due to its impressive capability in photo-realistic 3D reconstruction and novel view synthesis, given a set of posed camera images. Earlier work usually assumes the input images are of good quality. However, image degradation (e.g. image motion blur in low-light conditions) can easily happen in real-world scenarios, which would further affect the rendering quality of NeRF. In this paper, we present a novel bundle adjusted deblur Neural Radiance Fields (BAD-NeRF), which can be robust to severe motion blurred images and inaccurate camera poses. Our approach models the physical image formation process of a motion blurred image, and jointly learns the parameters of NeRF and recovers the camera motion trajectories during exposure time. In experiments, we show that by directly modeling the real physical image formation process, BAD-NeRF achieves superior performance over prior works on both synthetic and real datasets. Code and data are available at https://github.com/WU-CVGL/BAD-NeRF.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (58)
  1. RigNeRF: Fully Controllable Neural 3D Portraits. In Computer Vision and Pattern Recognition (CVPR), pages 20364–20373, 2022.
  2. Fast Motion Deblurring. ACM Transactions on Graphics (TOG), 28(5), 2009.
  3. Blender Online Community. Blender - a 3D modelling and rendering package. Blender Foundation, Stichting Blender Foundation, Amsterdam, 2018.
  4. 3D urban scene modeling integrating recognition and reconstruction. International Journal of Computer Vision (IJCV), 78(2-3):121–141, July 2008.
  5. Photometric Bundle Adjustment for Dense Multi-view 3D Modeling. In Computer Vision and Pattern Recognition (CVPR), 2014.
  6. GRAM: Generative Radiance Manifolds for 3D-Aware Image Generation. In Computer Vision and Pattern Recognition (CVPR), volume i, 2022.
  7. Removing camera shake from a single photograph. In ACM Transactions on Graphics (TOG), 2006.
  8. Accurate, Dense, and Robust Multi-View Stereopsis. IEEE Trans. Pattern Analysis and Machine Intelligence (PAMI), 32(8):1362–1376, 2010.
  9. Dynamic Neural Radiance Fields for Monocular 4D Facial Avatar Reconstruction. In Computer Vision and Pattern Recognition (CVPR). IEEE Computer Society, 2021.
  10. 3D Reconstruction Using an n-Layer Heightmap. In DAGM German Conference on Pattern Recognition (DAGM GCPR), 2010.
  11. FastNeRF: High-Fidelity Neural Rendering at 200FPS. In International Conference on Computer Vision (ICCV), 2021.
  12. StyleNeRF: A Style-Based 3D Aware Generator for High Resolution Image Synthesis. In International Conference on Learning Representations (ICLR), pages 1–20, 2022.
  13. Self-Calibrating Neural Radiance Fields. International Conference on Computer Vision (ICCV), 2021.
  14. Adam: A Method for Stochastic Optimization. arXiv preprint arXiv:1412.6980, 2014.
  15. Fast image deconvolution using Hyper-Laplacian priors. In Neural Information Processing Systems (NIPS), 2009.
  16. DeblurGAN-v2: Deblurring (Orders-of-Magnitude) Faster and Better. In International Conference on Computer Vision (ICCV), 2019.
  17. Pulsar: Efficient sphere-based neural rendering. In Computer Vision and Pattern Recognition (CVPR), 2021.
  18. Understanding and evaluating blind deconvolution algorithm. In Computer Vision and Pattern Recognition (CVPR), 2009.
  19. Marc Levoy. Efficient Ray Tracing of Volume Data. ACM Transactions on Graphics (TOG), 9(3):245–261, 1990.
  20. BARF: Bundle-Adjusting Neural Radiance Fields. In International Conference on Computer Vision (ICCV), 2021.
  21. MBA-VO: Motion Blur Aware Visual Odometry. In International Conference on Computer Vision (ICCV), pages 5550–5559, 2021.
  22. Editing Conditional Radiance Fields. In International Conference on Computer Vision (ICCV), 2021.
  23. Deblur-NeRF: Neural Radiance Fields from Blurry Images. In Computer Vision and Pattern Recognition (CVPR), pages 12861–12870, 2022.
  24. NeRF in the Wild: Neural Radiance Fields for Unconstrained Photo Collections. In Computer Vision and Pattern Recognition (CVPR), pages 7206–7215, 2021.
  25. Nelson Max. Optical Models for Direct Volume Rendering. IEEE Transactions on Visualization and Computer Graphics, 1(2):99–108, 1995.
  26. NeRF in the Dark: High Dynamic Range View Synthesis from Noisy Raw Images. In Computer Vision and Pattern Recognition (CVPR), pages 16190–16199, 2022.
  27. NeRF: Representing Scenes as Neural Radiance Fields for View Synthesis. In European Conference on Computer Vision (ECCV).
  28. Instant Neural Graphics Primitives with a Multiresolution Hash Encoding. ACM Transactions on Graphics (TOG), 41(4):102:1–102:15, July 2022.
  29. Deep Multi-Scale Convolutional Neural Network for Dynamic Scene Deblurring. In Computer Vision and Pattern Recognition (CVPR), July 2017.
  30. GIRAFFE: Representing Scenes as Compositional Generative Neural Feature Fields. In Computer Vision and Pattern Recognition (CVPR), pages 11448–11459, 2021.
  31. Real-time 3D Reconstruction at Scale using Voxel Hashing. In ACM Transactions on Graphics (TOG), 2013.
  32. StyleSDF: High-Resolution 3D-Consistent Image and Geometry Generation. In Computer Vision and Pattern Recognition (CVPR), pages 13503–13513, 2022.
  33. Joint Estimation of Camera Pose, Depth, Deblurring, and Super-Resolution from a Blurred Image Sequence. In International Conference on Computer Vision (ICCV), pages 4613–4621, 2017.
  34. Nerfies: Deformable Neural Radiance Fields. In International Conference on Computer Vision (ICCV), 2020.
  35. Animatable neural radiance fields for modeling dynamic human bodies. In International Conference on Computer Vision (ICCV), 2021.
  36. Neural Body: Implicit Neural Representations with Structured Latent Codes for Novel View Synthesis of Dynamic Humans. In Computer Vision and Pattern Recognition (CVPR), pages 9050–9059, 2021.
  37. TermiNeRF: Ray Termination Prediction for Efficient Neural Rendering. In International Conference on 3D Vision (3DV), 2021.
  38. D-NeRF: Neural Radiance Fields for Dynamic Scenes. In Computer Vision and Pattern Recognition (CVPR), pages 10313–10322, 2021.
  39. Derf: Decomposed radiance fields. In Computer Vision and Pattern Recognition (CVPR), pages 14148–14156, 2021.
  40. KiloNeRF: Speeding up Neural Radiance Fields with Thousands of Tiny MLPs. In International Conference on Computer Vision (ICCV), 2021.
  41. Structure-from-motion Revisited. In Computer Vision and Pattern Recognition (CVPR), pages 4104–4113, 2016.
  42. Bad slam: Bundle adjusted direct rgb-d slam. In Computer Vision and Pattern Recognition (CVPR), pages 134–144, 2019.
  43. High-quality Motion Deblurring from a Single Image. ACM Transactions on Graphics (TOG), 27(3), 2008.
  44. Recurrent Video Deblurring with Blur-Invariant Motion Estimation and Pixel Volumes. ACM Transactions on Graphics (TOG), 40(5):1–18, 2021.
  45. Volumetric 3D Mapping in Real-Time on a CPU. In International Conference on Robotics and Automation (ICRA), 2014.
  46. Deep Video Deblurring for Hand-held Cameras. In Computer Vision and Pattern Recognition (CVPR), 2017.
  47. Block-NeRF: Scalable Large Scene Neural View Synthesis. In Computer Vision and Pattern Recognition (CVPR), 2022.
  48. Scale-recurrent network for deep image deblurring. In Computer Vision and Pattern Recognition (CVPR), 2018.
  49. Mega-NeRF: Scalable Construction of Large-Scale NeRFs for Virtual Fly-Throughs. In Computer Vision and Pattern Recognition (CVPR), pages 12922–12931, 2022.
  50. CLIP-NeRF: Text-and-Image Driven Manipulation of Neural Radiance Fields. In Computer Vision and Pattern Recognition (CVPR), pages 1–10, 2022.
  51. HumanNeRF: Free-viewpoint Rendering of Moving People from Monocular Video. In Computer Vision and Pattern Recognition (CVPR), pages 16210–16220, 2022.
  52. ElasticFusion: Dense SLAM Without A Pose Graph. In Robotics: Science and Systems (RSS), 2015.
  53. NeX: Real-time View Synthesis with Neural Basis Expansion. In Computer Vision and Pattern Recognition (CVPR), pages 8530–8539, 2021.
  54. CityNeRF: Building NeRF at City Scale. In European Conference on Computer Vision (ECCV), 2022.
  55. Two-Phase Kernel Estimation for robust motion deblurring. In European Conference on Computer Vision (ECCV), 2010.
  56. PlenOctrees for Real-time Rendering of Neural Radiance Fields. In International Conference on Computer Vision (ICCV), 2021.
  57. Multi-stage Progressive Image Restoration. In Computer Vision and Pattern Recognition (CVPR), pages 14821–14831, 2021.
  58. The Unreasonable Effectiveness of Deep Features as a Perceptual Metric. In Computer Vision and Pattern Recognition (CVPR), pages 586–595, 2018.
Citations (51)

Summary

  • The paper introduces BAD-NeRF, which jointly optimizes NeRF parameters and camera motion trajectories to robustly handle motion blur in 3D scene reconstruction.
  • It employs a linear motion model in SE(3) to accurately capture exposure-time camera movements, directly integrating blur into the NeRF training pipeline.
  • Empirical results show improved PSNR, SSIM, and LPIPS metrics over existing methods, leading to better deblurring performance and novel view synthesis.

BAD-NeRF: Bundle Adjusted Deblur Neural Radiance Fields

The paper entitled "BAD-NeRF: Bundle Adjusted Deblur Neural Radiance Fields" presents an innovative solution to the problem of incorporating motion blurred images into Neural Radiance Fields (NeRF) frameworks, a topic of paramount interest in computer vision due to the growing application of NeRFs in 3D reconstruction and novel view synthesis. The researchers introduce the BAD-NeRF methodology, which jointly learns the parameters of NeRF and recovers camera motion trajectories during exposure time, offering a significant enhancement over previous methodologies that generally assume sharp input images.

Problem Context and Motivation

Traditional approaches in 3D scene reconstruction often rely on the assumption of high-quality, sharp images. The advent of NeRFs allowed for impressive 3D representations and view synthesis from multi-view posed images. However, scenarios with image degradations such as motion blur present a challenge for these approaches. Motion blur is especially prevalent in low-light conditions or high-speed camera motion, leading to degraded performance in NeRF training. This paper addresses the critical assumption violation within NeRF paradigms by proposing a framework that accounts for motion blur, enhancing training robustness and output quality.

Methodology

BAD-NeRF innovatively integrates the physical process of motion blur image formation into the NeRF training pipeline. The framework models the blur as a result of camera motion during exposure, and optimally estimates both camera motion trajectories and the NeRF parameters. Key to this approach is the linear motion model in SE(3) space for representing camera trajectory, which, within the typically short exposure times, assumes linear camera movement without compromising accuracy.

The methodological foundation of BAD-NeRF revolves around:

  • Joint Optimization: Simultaneously optimizing the NeRF parameters and camera poses during the exposure period, overcoming limitations from inaccurate initial camera position estimates.
  • Exposure-Time Motion Modeling: Utilizing a linear model to interpolate poses in SE(3) space, this formulation captures the trajectory of the camera during the motion blur, facilitating accurate rendering and synthesis.

The articulated photo-metric bundle adjustment formulation under the NeRF framework allows a direct simulation of the image formation process with motion blur, advancing state-of-the-art performance in both deblurring and synthesis tasks.

Empirical Evaluation and Results

The empirical assessment conducted on both synthetic datasets and real-world representations from prior research, including datasets from Deblur-NeRF and MBA-VO, underscores the efficacy of BAD-NeRF. Quantitative metrics such as PSNR, SSIM, and LPIPS demonstrate superior performance over several contemporary deblurring techniques, both single image and multi-view based.

  • Deblurring Performance: BAD-NeRF significantly outperforms rivals like SRNDeblurNet and Deblur-NeRF, especially in scenarios with severe image motion blur. Notably, BAD-NeRF maintains robust performance regardless of the blur level, highlighting its adaptability.
  • Novel View Synthesis: Achieving high-quality novel view renderings with better image metrics by addressing occlusions effectively.
  • Pose Estimation: Improved accuracy in camera pose estimation compared to methods such as BARF, facilitated by explicit trajectory modeling.

Implications and Future Directions

The practical implications of BAD-NeRF are substantial, providing a reliable framework for applications requiring precise 3D rendering from degraded image sets, such as autonomous navigation and augmented reality. Theoretically, this work paves the way for further enhancements in NeRF robustness and versatility, including potential integrations with dynamic scene or non-rigid object reconstructions.

Future advancements could explore more complex camera motion models or adaptive time-dependent trajectories to capture even more intricate movements, possibly extending BAD-NeRF’s utility in more varied environments and broader deployment scenarios. Additionally, exploring integration with other visual SLAM and tracking systems could further enhance its practical application.

In conclusion, BAD-NeRF charts a forward path in the handling of motion blurred images within NeRF systems, effectively bridging the gap between theoretical robustness and practical applicability, and opening myriad possibilities for both academic inquiry and real-world deployment.

Github Logo Streamline Icon: https://streamlinehq.com