Collaborative Dynamic 3D Scene Graphs for Automated Driving (2309.06635v3)
Abstract: Maps have played an indispensable role in enabling safe and automated driving. Although there have been many advances on different fronts ranging from SLAM to semantics, building an actionable hierarchical semantic representation of urban dynamic scenes and processing information from multiple agents are still challenging problems. In this work, we present Collaborative URBan Scene Graphs (CURB-SG) that enable higher-order reasoning and efficient querying for many functions of automated driving. CURB-SG leverages panoptic LiDAR data from multiple agents to build large-scale maps using an effective graph-based collaborative SLAM approach that detects inter-agent loop closures. To semantically decompose the obtained 3D map, we build a lane graph from the paths of ego agents and their panoptic observations of other vehicles. Based on the connectivity of the lane graph, we segregate the environment into intersecting and non-intersecting road areas. Subsequently, we construct a multi-layered scene graph that includes lane information, the position of static landmarks and their assignment to certain map sections, other vehicles observed by the ego agents, and the pose graph from SLAM including 3D panoptic point clouds. We extensively evaluate CURB-SG in urban scenarios using a photorealistic simulator. We release our code at http://curb.cs.uni-freiburg.de.
- B. Yang, M. Liang, and R. Urtasun, “HDNET: Exploiting HD maps for 3D object detection,” in Conf. on Robot Learning, 2018.
- D. Cattaneo, D. G. Sorrenti, and A. Valada, “CMRNet++: Map and camera agnostic monocular visual localization in LiDAR maps,” Int. Conf. on Robotics and Automation Workshop on Emerging Learning and Alg. Methods for Data Association in Robotics, 2020.
- A. Diaz-Diaz, M. Ocaña, A. Llamazares, C. Gómez-Huélamo, P. Revenga, and L. M. Bergasa, “HD maps: Exploiting opendrive potential for path planning and map monitoring,” in IEEE Intelligent Vehicles Symposium, 2022, pp. 1211–1217.
- R. Trumpp, M. Büchner, A. Valada, and M. Caccamo, “Efficient learning of urban driving policies using bird’s-eye-view state representations,” Int. Conf. on Intelligent Transportation Systems, 2023.
- F. Poggenhans, J.-H. Pauls, J. Janosovits, S. Orf, M. Naumann, F. Kuhnt, and M. Mayr, “Lanelet2: A high-definition map framework for the future of automated driving,” in Int. Conf. on Intelligent Transportation Systems, 2018.
- X. Chen, A. Milioto, E. Palazzolo, P. Giguère, J. Behley, and C. Stachniss, “SuMa++: Efficient LiDAR-based semantic SLAM,” in Int. Conf. on Intelligent Robots and Systems, 2019, pp. 4530–4537.
- K. Koide, J. Miura, and E. Menegatti, “A portable three-dimensional LiDAR-based system for long-term and wide-area people behavior measurement,” Int. Journal of Adv. Rob. Systems, vol. 16, no. 2, 2019.
- K. Wong, Y. Gu, and S. Kamijo, “Mapping for autonomous driving: Opportunities and challenges,” IEEE Intelligent Transportation Systems Magazine, vol. 13, no. 1, pp. 91–106, 2020.
- A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark, et al., “Learning transferable visual models from natural language supervision,” in Int. Conf. on Machine Learning. PMLR, 2021, pp. 8748–8763.
- H. Bavle, J. L. Sanchez-Lopez, M. Shaheer, J. Civera, and H. Voos, “S-Graphs+: Real-time localization and mapping leveraging hierarchical representations,” IEEE Robotics and Automation Letters, vol. 8, no. 8, pp. 4927–4934, 2023.
- N. Hughes, Y. Chang, and L. Carlone, “Hydra: A real-time spatial perception system for 3D scene graph construction and optimization,” in Robotics: Science and Systems, 2022.
- I. Armeni, Z.-Y. He, A. Zamir, J. Gwak, J. Malik, M. Fischer, and S. Savarese, “3D scene graph: A structure for unified semantics, 3D space, and camera,” in Int. Conf. on Computer Vision, 2019, pp. 5663–5672.
- A. Dosovitskiy, G. Ros, F. Codevilla, A. Lopez, and V. Koltun, “CARLA: An open urban driving simulator,” in Conf. on Robot Learning, 2017.
- J. Zhang and S. Singh, “LOAM: Lidar odometry and mapping in real-time.” in Robotics: Science and Systems, 2014.
- J. Arce, N. Vödisch, D. Cattaneo, W. Burgard, and A. Valada, “PADLoC: LiDAR-based deep loop closure detection and registration using panoptic attention,” IEEE Robotics and Automation Letters, vol. 8, no. 3, pp. 1319–1326, 2023.
- D. Zou, P. Tan, and W. Yu, “Collaborative visual SLAM for multiple agents: A brief survey,” Virtual Reality and Intelligent Hardware, vol. 1, no. 5, pp. 461–482, 2019.
- L. Riazuelo, J. Civera, and J. Montiel, “C2TAM: A cloud framework for cooperative tracking and mapping,” Robotics and Autonomous Systems, vol. 62, no. 4, pp. 401–413, 2014.
- M. Karrer, P. Schmuck, and M. Chli, “CVI-SLAM — collaborative visual-inertial SLAM,” IEEE Robotics and Automation Letters, vol. 3, no. 4, pp. 2762–2769, 2018.
- P. Schmuck, T. Ziegler, M. Karrer, J. Perraudin, and M. Chli, “COVINS: Visual-inertial SLAM for centralized collaboration,” in IEEE Int. Symp. on Mixed and Augmented Reality Adjunct, 2021, pp. 171–176.
- Y. Chang, K. Ebadi, C. E. Denniston, M. F. Ginting, A. Rosinol, A. Reinke, et al., “LAMP 2.0: A robust multi-robot SLAM system for operation in challenging large-scale underground environments,” IEEE Robotics and Automation Letters, vol. 7, no. 4, pp. 9175–9182, 2022.
- P.-Y. Lajoie and G. Beltrame, “Swarm-SLAM: Sparse decentralized collaborative simultaneous localization and mapping framework for multi-robot systems,” arXiv preprint arXiv:2301.06230, 2023.
- Y. Tian, Y. Chang, F. Herrera Arias, C. Nieto-Granda, J. P. How, and L. Carlone, “Kimera-Multi: Robust, distributed, dense metric-semantic SLAM for multi-robot systems,” IEEE Trans. on Robotics, vol. 38, no. 4, pp. 2022–2038, 2022.
- Y. Huang, T. Shan, F. Chen, and B. Englot, “DiSCo-SLAM: Distributed scan context-enabled multi-robot LiDAR SLAM with two-stage global-local graph optimization,” IEEE Robotics and Automation Letters, vol. 7, no. 2, pp. 1150–1157, 2022.
- G. Kim and A. Kim, “Scan Context: Egocentric spatial descriptor for place recognition within 3D point cloud map,” in Int. Conf. on Intelligent Robots and Systems, 2018, pp. 4802–4809.
- H. Bavle, J. L. Sanchez-Lopez, M. Shaheer, J. Civera, and H. Voos, “Situational graphs for robot navigation in structured indoor environments,” IEEE Robotics and Automation Letters, vol. 7, no. 4, pp. 9107–9114, 2022.
- C. Lang, A. Braun, and A. Valada, “Robust object detection using knowledge graph embeddings,” in DAGM German Conference on Pattern Recognition, 2022, pp. 445–461.
- M. Büchner, J. Zürn, I.-G. Todoran, A. Valada, and W. Burgard, “Learning and aggregating lane graphs for urban automated driving,” in IEEE/CVF Conf. on Computer Vision and Pattern Recognition, 2023, pp. 13 415–13 424.
- Liao, Bencheng and Chen, Shaoyu and Wang, Xinggang and Cheng, Tianheng, and Zhang, Qian and Liu, Wenyu and Huang, Chang, “MapTR: Structured modeling and learning for online vectorized HD map construction,” in Int. Conf. on Learning Representations, 2023.
- S. He and H. Balakrishnan, “Lane-level street map extraction from aerial imagery,” in IEEE Winter Conference on Applications of Computer Vision, January 2022, pp. 2080–2089.
- N. Gosala, K. Petek, P. L. Drews-Jr, W. Burgard, and A. Valada, “Skyeye: Self-supervised bird’s-eye-view semantic mapping using monocular frontal view images,” in IEEE/CVF Conf. on Computer Vision and Pattern Recognition, 2023, pp. 14 901–14 910.
- R. Mohan and A. Valada, “Amodal panoptic segmentation,” in IEEE/CVF Conf. on Computer Vision and Pattern Recognition, 2022, pp. 21 023–21 032.
- R. Krishna, Y. Zhu, O. Groth, J. Johnson, K. Hata, J. Kravitz, et al., “Visual genome: Connecting language and vision using crowdsourced dense image annotations,” Int. Journal of Computer Vision, vol. 123, pp. 32–73, 2017.
- A. Rosinol, A. Gupta, M. Abate, J. Shi, and L. Carlone, “3D dynamic scene graphs: Actionable spatial perception with places, objects, and humans,” Robotics: Science and Systems, 2020.
- U.-H. Kim, J.-M. Park, T.-j. Song, and J.-H. Kim, “3-D scene graph: A sparse and semantic representation of physical environments for intelligent agents,” IEEE Trans. on Cybernetics, vol. 50, no. 12, pp. 4921–4933, 2020.
- J. Johnson, B. Hariharan, L. Van Der Maaten, J. Hoffman, L. Fei-Fei, C. Lawrence Zitnick, and R. Girshick, “Inferring and executing programs for visual reasoning,” in Int. Conf. on Computer Vision, 2017, pp. 2989–2998.
- J. Wald, H. Dhamo, N. Navab, and F. Tombari, “Learning 3d semantic scene graphs from 3d indoor reconstructions,” in IEEE/CVF Conf. on Computer Vision and Pattern Recognition, 2020.
- S.-C. Wu, J. Wald, K. Tateno, N. Navab, and F. Tombari, “SceneGraphFusion: Incremental 3D scene graph prediction from RGB-D sequences,” in IEEE/CVF Conf. on Computer Vision and Pattern Recognition, June 2021, pp. 7515–7525.
- K. Sirohi, R. Mohan, D. Büscher, W. Burgard, and A. Valada, “EfficientLPS: Efficient LiDAR panoptic segmentation,” IEEE Trans. on Robotics, vol. 38, no. 3, pp. 1894–1914, 2022.
- K. Koide, M. Yokozuka, S. Oishi, and A. Banno, “Voxelized GICP for fast and accurate 3D point cloud registration,” in Int. Conf. on Robotics and Automation, 2021, pp. 11 054–11 059.
- R. Kümmerle, G. Grisetti, H. Strasdat, K. Konolige, and W. Burgard, “g2o: A general framework for graph optimization,” in Int. Conf. on Robotics and Automation, 2011, pp. 3607–3613.
- M. Ester, H.-P. Kriegel, J. Sander, X. Xu, et al., “A density-based algorithm for discovering clusters in large spatial databases with noise,” in ACM SIGKDD Conf. on Knowledge Discovery and Data Mining, vol. 96, no. 34, 1996, pp. 226–231.
- G. Jozkow, P. Wieczorek, M. Karpina, A. Walicka, and A. Borkowski, “Performance evaluation of sUAS equipped with Velodyne HDL-32E LiDAR sensor,” The Int. Archives of the Photogrammetry, Remote Sensing and Spatial Information Sciences, vol. 42, pp. 171–177, 2017.
- A. Van, D. Lindenbaum, and T. M. Bacastow, “SpaceNet: A remote sensing dataset and challenge series,” arXiv preprint arXiv:1807.01232, 2018.
Sponsored by Paperpile, the PDF & BibTeX manager trusted by top AI labs.
Get 30 days freePaper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.