Learning to Robustly Reconstruct Low-light Dynamic Scenes from Spike Streams (2401.10461v2)
Abstract: As a neuromorphic sensor with high temporal resolution, spike camera can generate continuous binary spike streams to capture per-pixel light intensity. We can use reconstruction methods to restore scene details in high-speed scenarios. However, due to limited information in spike streams, low-light scenes are difficult to effectively reconstruct. In this paper, we propose a bidirectional recurrent-based reconstruction framework, including a Light-Robust Representation (LR-Rep) and a fusion module, to better handle such extreme conditions. LR-Rep is designed to aggregate temporal information in spike streams, and a fusion module is utilized to extract temporal features. Additionally, we have developed a reconstruction benchmark for high-speed low-light scenes. Light sources in the scenes are carefully aligned to real-world conditions. Experimental results demonstrate the superiority of our method, which also generalizes well to real spike streams. Related codes and proposed datasets will be released after publication.
- A 240 × 180 130 db 3 μs𝜇s\mu\rm sitalic_μ roman_s latency global shutter spatiotemporal vision sensor. IEEE Journal of Solid-State Circuits (JSSC), 49(10):2333–2341, 2014.
- Self-supervised mutual learning for dynamic scene reconstruction of spiking camera. In Proceedings of the International Joint Conference on Artificial Intelligence, IJCAI, pages 2859–2866, 2022.
- Activity-driven, event-based vision sensors. IEEE International Symposium on Circuits and Systems (ISCAS), pages 2426–2429, 2010.
- High-speed scene reconstruction from low-light spike streams. In 2022 IEEE International Conference on Visual Communications and Image Processing (VCIP), pages 1–5. IEEE, 2022.
- Optimal biasing and physical limits of dvs event noise. arXiv preprint arXiv:2304.04019, 2023.
- Zero-reference deep curve estimation for low-light image enhancement. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, pages 1780–1789, 2020.
- Optical flow estimation for spiking camera. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 17844–17853, 2022.
- 1000×\times× faster camera and machine vision with ordinary devices. Engineering, 2022.
- Enlightengan: Deep light enhancement without paired supervision. IEEE transactions on image processing, 30:2340–2349, 2021.
- Retinomorphic sensing: A novel paradigm for future multimedia computing. In Proceedings of the ACM International Conference on Multimedia (ACMMM), page 144–152, 2021.
- Low-light image and video enhancement using deep learning: A survey. IEEE transactions on pattern analysis and machine intelligence, 44(12):9396–9416, 2021.
- Learning to enhance low-light image via zero-reference deep curve estimation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 44(8):4225–4238, 2021.
- Photon-limited object detection using non-local feature matching and knowledge distillation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 3976–3987, 2021.
- Uncertainty guided depth fusion for spike camera. arXiv preprint arXiv:2208.12653, 2022.
- A 128 × 128 120 db 15 μs𝜇s\mu\rm sitalic_μ roman_s latency asynchronous temporal contrast vision sensor. IEEE Journal of Solid-State Circuits (JSSC), 43(2):566–576, 2008.
- Unsupervised spike depth estimation via cross-modality cross-domain knowledge transfer. arXiv preprint arXiv:2208.12527, 2022.
- Yolo in the dark-domain adaptation method for merging multiple models. In Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XXI 16, pages 345–359. Springer, 2020.
- Research on improved yolov5 for low-light environment object detection. Electronics, 12(14):3089, 2023.
- Edvr: Video restoration with enhanced deformable convolutional networks. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR) Workshops, June 2019.
- Deep retinex decomposition for low-light enhancement. arXiv preprint arXiv:1808.04560, 2018.
- Tom D Wilson. On user studies and information needs. Journal of documentation, 37(1):3–15, 1981.
- A bio-inspired multi-exposure fusion framework for low-light image enhancement. arXiv preprint arXiv:1711.00591, 2017.
- Spike-based optical flow estimation via contrastive learning. In ICASSP 2023-2023 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pages 1–5. IEEE, 2023.
- Learning temporal-ordered representation for spike streams based on discrete wavelet transforms. In Proceedings of the AAAI Conference on Artificial Intelligence, volume 37, pages 137–147, 2023.
- Spike transformer: Monocular depth estimation for spiking camera. In European Conference on Computer Vision (ECCV), 2022.
- High-speed motion scene reconstruction for spike camera via motion aligned filtering. In International Symposium on Circuits and Systems (ISCAS), pages 1–5, 2020.
- Spk2imgnet: Learning to reconstruct dynamic scene from continuous spike stream. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 11996–12005, 2021.
- Spikingsim: A bio-inspired spiking simulator. In 2022 IEEE International Symposium on Circuits and Systems (ISCAS), pages 3003–3007. IEEE, 2022.
- Learninng optical flow from continuous spike streams. In Proceedings of the Annual Conference on Neural Information Processing Systems (NeurIPS), 2022.
- High-speed image reconstruction through short-term plasticity for spiking cameras. In IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pages 6358–6367, 2021.
- A retina-inspired sampling method for visual texture reconstruction. In IEEE International Conference on Multimedia and Expo (ICME), pages 1432–1437, 2019.
- Ultra-high temporal resolution visual reconstruction from a fovea-like spike camera via spiking neuron model. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(1):1233–1249, 2022.
- Neuspike-net: High speed video reconstruction via bio-inspired neuromorphic cameras. In IEEE International Conference on Computer Vision (ICCV), pages 2400–2409, 2021.