Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
157 tokens/sec
GPT-4o
43 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

SpikeNeRF: Learning Neural Radiance Fields from Continuous Spike Stream (2403.11222v1)

Published 17 Mar 2024 in cs.CV

Abstract: Spike cameras, leveraging spike-based integration sampling and high temporal resolution, offer distinct advantages over standard cameras. However, existing approaches reliant on spike cameras often assume optimal illumination, a condition frequently unmet in real-world scenarios. To address this, we introduce SpikeNeRF, the first work that derives a NeRF-based volumetric scene representation from spike camera data. Our approach leverages NeRF's multi-view consistency to establish robust self-supervision, effectively eliminating erroneous measurements and uncovering coherent structures within exceedingly noisy input amidst diverse real-world illumination scenarios. The framework comprises two core elements: a spike generation model incorporating an integrate-and-fire neuron layer and parameters accounting for non-idealities, such as threshold variation, and a spike rendering loss capable of generalizing across varying illumination conditions. We describe how to effectively optimize neural radiance fields to render photorealistic novel views from the novel continuous spike stream, demonstrating advantages over other vision sensors in certain scenes. Empirical evaluations conducted on both real and novel realistically simulated sequences affirm the efficacy of our methodology. The dataset and source code are released at https://github.com/BIT-Vision/SpikeNeRF.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (51)
  1. Mip-nerf: A multiscale representation for anti-aliasing neural radiance fields. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 5855–5864, 2021.
  2. A 240 ×\times× 180 130 d⁢b𝑑𝑏dbitalic_d italic_b 3 μ⁢s𝜇𝑠\mu sitalic_μ italic_s latency global shutter spatiotemporal vision sensor. IEEE J. Solid-State Circuits, 49(10):2333–2341, 2014.
  3. Self-supervised mutual learning for dynamic scene reconstruction of spiking camera. In Proceedings of the Thirty-First International Joint Conference on Artificial Intelligence, IJCAI-22, pages 2859–2866. International Joint Conferences on Artificial Intelligence Organization, 2022. Main Track.
  4. Activity-driven, event-based vision sensors. In IEEE Int. Symp. Circuits Syst. (ISCAS), pages 2426–2429, 2010.
  5. Depth-supervised nerf: Fewer views and faster training for free. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12882–12891, 2022.
  6. Spike coding for dynamic vision sensor in intelligent driving. IEEE Internet of Things Journal, 6(1):60–71, 2018.
  7. An efficient coding method for spike camera using inter-spike intervals. In 2019 Data Compression Conference (DCC), pages 568–568, 2019.
  8. Fastnerf: High-fidelity neural rendering at 200fps. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 14346–14355, 2021.
  9. E-raft: Dense optical flow from event cameras. In 2021 International Conference on 3D Vision (3DV), pages 197–206. IEEE, 2021.
  10. Neuronal dynamics: From single neurons to networks and models of cognition. Cambridge University Press, 2014.
  11. Reliable event generation with invertible conditional normalizing flow. IEEE Transactions on Pattern Analysis and Machine Intelligence, 2023.
  12. Neuromorphic camera guided high dynamic range imaging. In IEEE Conf. Comput. Vis. Pattern Recog. (CVPR), 2020.
  13. Learning monocular dense depth from events. In 2020 International Conference on 3D Vision (3DV), pages 534–542. IEEE, 2020.
  14. Optical flow estimation for spiking camera. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 17844–17853, 2022.
  15. 1000×\times× faster camera and machine vision with ordinary devices. Engineering, 25:110–119, 2023.
  16. Hdr-nerf: High dynamic range neural radiance fields. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 18398–18408, 2022.
  17. Ev-nerf: Event based neural radiance field. pages 837–847, 2023.
  18. Neural scene flow fields for space-time view synthesis of dynamic scenes. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 6498–6508, 2021.
  19. Robust e-nerf: Nerf from sparse & noisy events under non-uniform motion. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 18335–18346, 2023.
  20. Deblur-nerf: Neural radiance fields from blurry images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12861–12870, 2022.
  21. Nerf in the wild: Neural radiance fields for unconstrained photo collections. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 7210–7219, 2021.
  22. Nerf: Representing scenes as neural radiance fields for view synthesis. Communications of the ACM, 65(1):99–106, 2021.
  23. Nerf in the dark: High dynamic range view synthesis from noisy raw images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 16190–16199, 2022.
  24. No-reference image quality assessment in the spatial domain. IEEE Transactions on image processing, 21(12):4695–4708, 2012a.
  25. Making a “completely blind” image quality analyzer. IEEE Signal processing letters, 20(3):209–212, 2012b.
  26. Surrogate gradient learning in spiking neural networks: Bringing the power of gradient-based optimization to spiking neural networks. IEEE Signal Processing Magazine, 36(6):51–63, 2019.
  27. Regnerf: Regularizing neural radiance fields for view synthesis from sparse inputs. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 5480–5490, 2022.
  28. E2nerf: Event enhanced neural radiance fields from blurry images. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 13254–13264, 2023.
  29. Eventnerf: Neural radiance fields from a single colour event camera. arXiv preprint arXiv:2206.11896, 2022.
  30. Front and back illuminated dynamic and active pixel vision sensors comparison. IEEE Transactions on Circuits and Systems II: Express Briefs, 65(5):677–681, 2018.
  31. Time lens: Event-based video frame interpolation. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 16155–16164, 2021.
  32. Visevent: Reliable object tracking via collaboration of frame and event flows. IEEE Transactions on Cybernetics, 2023.
  33. Learning stereo depth estimation with bio-inspired spike cameras. In 2022 IEEE International Conference on Multimedia and Expo (ICME), pages 1–6. IEEE, 2022.
  34. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing, 13(4):600–612, 2004.
  35. Paul J Werbos. Backpropagation through time: what it does and how to do it. Proceedings of the IEEE, 78(10):1550–1560, 1990.
  36. Progressive tandem learning for pattern recognition with deep spiking neural networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(11):7824–7840, 2021.
  37. Learning super-resolution reconstruction for high temporal resolution spike stream. IEEE Transactions on Circuits and Systems for Video Technology, 33(1):16–29, 2021.
  38. Eventcap: Monocular 3d capture of high-speed human motions using an event camera. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4968–4978, 2020.
  39. pixelnerf: Neural radiance fields from one or few images. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 4578–4587, 2021.
  40. A highly effective and robust membrane potential-driven supervised learning method for spiking neurons. IEEE transactions on neural networks and learning systems, 30(1):123–137, 2018a.
  41. Rectified linear postsynaptic potential function for backpropagation in deep spiking neural networks. IEEE transactions on neural networks and learning systems, 33(5):1947–1958, 2021.
  42. The unreasonable effectiveness of deep features as a perceptual metric. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 586–595, 2018b.
  43. High-speed motion scene reconstruction for spike camera via motion aligned filtering. In IEEE Int. Symp. Circuits Syst. (ISCAS), pages 1–5, 2020.
  44. Super resolve dynamic scene from continuous spike streams. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 2533–2542, 2021a.
  45. Spk2imgnet: Learning to reconstruct dynamic scene from continuous spike stream. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 11996–12005, 2021b.
  46. A retina-inspired sampling method for visual texture reconstruction. In IEEE Int. Conf. Multimedia Expo. (ICME), pages 1432–1437, 2019.
  47. Hybrid coding of spatiotemporal spike data for a bio-inspired camera. IEEE Trans. on Circuits and Systems for Video Technology, 2020a.
  48. Retina-like visual image reconstruction via spiking neural model. In IEEE Conf. Comput. Vis. Pattern Recog. (CVPR), pages 1438–1446, 2020b.
  49. Neuspike-net: High speed video reconstruction via bio-inspired neuromorphic cameras. In IEEE Int. Conf. Comput. Vis. (ICCV), pages 2400–2409, 2021.
  50. Ultra-high temporal resolution visual reconstruction from a fovea-like spike camera via spiking neuron model. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(1):1233–1249, 2022.
  51. Recurrent spike-based image restoration under general illumination. In Proceedings of the 31st ACM International Conference on Multimedia, pages 8251–8260, 2023.
Citations (2)

Summary

We haven't generated a summary for this paper yet.