LaserSAM: Zero-Shot Change Detection Using Visual Segmentation of Spinning LiDAR (2402.10321v2)
Abstract: This paper presents an approach for applying camera perception techniques to spinning LiDAR data. To improve the robustness of long-term change detection from a 3D LiDAR, range and intensity information are rendered into virtual perspectives using a pinhole camera model. Hue-saturation-value image encoding is used to colourize the images by range and near-IR intensity. The LiDAR's active scene illumination makes it invariant to ambient brightness, which enables night-to-day change detection without additional processing. Using the range-colourized, perspective image allows existing foundation models to detect semantic regions. Specifically, the Segment Anything Model detects semantically similar regions in both a previously acquired map and live view from a path-repeating robot. By comparing the masks in both views, changes in the live scan are detected. Results indicate that the Segment Anything Model accurately captures the shape of arbitrary changes introduced into scenes. The proposed method achieves a segmentation intersection over union of 73.3% when evaluated in unstructured environments and 80.4% when evaluated within the planning corridor. Changes can be detected reliably through day-to-night illumination variations. After pixel-level masks are generated, the one-to-one correspondence with 3D points means that the 2D masks can be used directly to recover the 3D location of the changes. The detected 3D changes are avoided in a closed loop by treating them as obstacles in a local motion planner. Experiments on an unmanned ground vehicle demonstrate the performance of the method.
- Z. Ma, Y. Yang, G. Wang, X. Xu, H. T. Shen, and M. Zhang, “Rethinking Open-World Object Detection in Autonomous Driving Scenarios,” in Proceedings of the 30th ACM International Conf. on Multimedia, New York, NY, USA, Oct. 2022, pp. 1279–1288.
- R. Qin, J. Tian, and P. Reinartz, “3d change detection – approaches and applications,” ISPRS J. of Photogrammetry and Remote Sensing, vol. 122, pp. 41–56, 2016.
- P. Papadakis, “Terrain traversability analysis methods for unmanned ground vehicles: A survey,” Engineering Applications of Artificial Intelligence, vol. 26, no. 4, pp. 1373–1385, 2013.
- P. Furgale and T. D. Barfoot, “Visual teach and repeat for long-range rover autonomy,” J. of Field Robotics, 2010.
- L.-P. Berczi and T. D. Barfoot, “It’s like déjà vu all over again: Learning place-dependent terrain assessment for visual teach and repeat,” in 2016 IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2016, pp. 3973–3980.
- Y. Wu, “VT&R3: Generalizing the teach and repeat navigation framework,” Sep. 2022, mASc Thesis.
- J. Sehn, J. Collier, and T. D. Barfoot, “Off the beaten track: Laterally weighted motion planning for local obstacle avoidance,” 2023.
- C. Stauffer and W. Grimson, “Learning patterns of activity using real-time tracking,” IEEE Trans. Pattern Anal. Mach. Intell., vol. 22, no. 8, pp. 747–757, 2000.
- U. Okyay, J. Telling, C. L. Glennie, and W. E. Dietrich, “Airborne lidar change detection: An overview of earth sciences applications,” Earth-Science Reviews, vol. 198, p. 102929, 2019.
- L. Ding, K. Zhu, D. Peng, H. Tang, K. Yang, and B. Lorenzo, “Adapting segment anything model for change detection in hr remote sensing images,” arXiv preprint arXiv:2309.01429, 2023.
- I. de Gélis, S. Lefèvre, and T. Corpetti, “Siamese KPConv: 3D multiple change detection from raw point clouds using deep learning,” ISPRS J. of Photogrammetry and Remote Sensing, vol. 197, pp. 274–291, Mar. 2023.
- A. Krawciw, J. Sehn, and T. D. Barfoot, “Change of scenery: Unsupervised lidar change detection for mobile robots,” 2023.
- Y. Chen, B. Xu, F. Dümbgen, and T. D. B. Barfoot, “What to learn: Features, image transformations, or both?” in IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS). IEEE, 2023.
- L. Wijayathunga, A. Rassau, and D. Chai, “Challenges and solutions for autonomous ground robot scene understanding and navigation in unstructured outdoor environments: A review,” Applied Sciences, vol. 13, no. 17, p. 9877, 2023, number: 17 MDPI.
- Ouster. How multi-beam flash lidar works. [Online]. Available: https://ouster.com/insights/blog/how-multi-beam-flash-lidar-works
- A. Kirillov, E. Mintun, N. Ravi, H. Mao, C. Rolland, L. Gustafson, T. Xiao, S. Whitehead, A. C. Berg, W.-Y. Lo, P. Dollár, and R. Girshick, “Segment anything,” 2023.
- X. Zhao, W. Ding, Y. An, Y. Du, T. Yu, M. Li, M. Tang, and J. Wang, “Fast Segment Anything,” 2023.
- D. Girardeau-Montaut, M. Roux, R. Marc, and G. Thibault, “Change detection on points cloud data acquired with a ground laser scanner,” International Archives of Photogrammetry, Remote Sensing and Spatial Information Sciences, vol. 36, no. 3, p. W19, 2005.
- J. P. Underwood, D. Gillsjö, T. Bailey, and V. Vlaskine, “Explicit 3D change detection using ray-tracing in spherical coordinates,” in 2013 IEEE International Conf. on Robotics and Automation, May 2013, pp. 4735–4741, iSSN: 1050-4729.
- I. de Gélis, S. Saha, M. Shahzad, T. Corpetti, S. Lefèvre, and X. Zhu, “Deep unsupervised learning for 3d als point clouds change detection,” ISPRS Open Journal of Photogrammetry and Remote Sensing, vol. 9, p. 100044, 08 2023.
- Y. Gao, H. Yuan, T. Ku, R. C. Veltkamp, G. Zamanakos, L. Tsochatzidis, A. Amanatiadis, I. Pratikakis, A. Panou, I. Romanelis, V. Fotis, G. Arvanitis, and K. Moustakas, “SHREC 2023: Point cloud change detection for city scenes,” Computers & Graphics, vol. 115, pp. 35–42, 2023.
- P.-L. St-Charles, G.-A. Bilodeau, and R. Bergevin, “Subsense: A universal change detection method with local adaptive sensitivity,” IEEE Transactions on Image Processing, vol. 24, no. 1, pp. 359–373, 2015.
- C. Stauffer and W. Grimson, “Learning patterns of activity using real-time tracking,” IEEE Transactions on Pattern Analysis and Machine Intelligence, vol. 22, no. 8, pp. 747–757, 2000.
- K. Toyama, J. Krumm, B. Brumitt, and B. Meyers, “Wallflower: principles and practice of background maintenance,” in Proceedings of the Seventh IEEE International Conf. on Computer Vision, vol. 1, 1999, pp. 255–261 vol.1.
- M. Cordts, M. Omran, S. Ramos, T. Rehfeld, M. Enzweiler, R. Benenson, U. Franke, S. Roth, and B. Schiele, “The cityscapes dataset for semantic urban scene understanding,” in Proc. of the IEEE Conf. on Computer Vision and Pattern Recognition (CVPR), 2016.
- J. Behley, M. Garbade, A. Milioto, J. Quenzel, S. Behnke, C. Stachniss, and J. Gall, “SemanticKITTI: A Dataset for Semantic Scene Understanding of LiDAR Sequences,” in Proc. of the IEEE/CVF International Conf. on Computer Vision (ICCV), 2019.
- “Ouster OS1 LiDAR,” Available Online [https://ouster.com/products/scanning-lidar/os1-sensor/].
- “Robosense LiDAR,” Available Online [https://www.robosense.ai/en].
- C. McManus, P. Furgale, and T. D. Barfoot, “Towards lighting-invariant visual navigation: An appearance-based approach using scanning laser-rangefinders,” Robotics and Autonomous Systems, vol. 61, no. 8, pp. 836–852, 2013.
- S. Anderson and T. D. Barfoot, “Ransac for motion-distorted 3d visual sensors,” in 2013 IEEE/RSJ International Conf. on Intelligent Robots and Systems. IEEE, 2013, pp. 2093–2099.
- C. McManus, P. Furgale, B. Stenning, and T. D. Barfoot, “Lighting-invariant visual teach and repeat using appearance-based lidar,” J. of Field Robotics, vol. 30, no. 2, pp. 254–287, 2013.
- T. D. Barfoot, C. McManus, S. R. Anderson, H. Dong, E. Beerepoot, C. H. Tong, P. T. Furgale, J. D. Gammell, and J. Enright, “Into darkness: Visual navigation based on a lidar-intensity-image pipeline,” in International Symposium of Robotics Research, 2013.
- K. Burnett, Y. Wu, D. J. Yoon, A. P. Schoellig, and T. D. Barfoot, “Are We Ready for Radar to Replace Lidar in All-Weather Mapping and Localization?” IEEE Robotics and Automation Letters, vol. 7, no. 4, pp. 10 328–10 335, Oct. 2022.
- “Clearpath Robotics Warthog UGV,” 2020. [Online]. Available: https://clearpathrobotics.com/warthog-unmanned-ground-vehicle-robot/
- C. J. Van Rijsbergen, “Chapter 7: Evaluation,” in Information Retrieval, 2nd ed. Butterworths.