Research on Reliable and Safe Occupancy Grid Prediction in Underground Parking Lots (2407.02197v1)
Abstract: Against the backdrop of advancing science and technology, autonomous vehicle technology has emerged as a focal point of intense scrutiny within the academic community. Nevertheless, the challenge persists in guaranteeing the safety and reliability of this technology when navigating intricate scenarios. While a substantial portion of autonomous driving research is dedicated to testing in open-air environments, such as urban roads and highways, where the myriad variables at play are meticulously examined, enclosed indoor spaces like underground parking lots have, to a significant extent, been overlooked in the scholarly discourse. This discrepancy highlights a gap in derstanding the unique challenges these confined settings pose for autonomous navigation systems. This study tackles indoor autonomous driving, particularly in overlooked spaces like underground parking lots. Using CARLA's simulation platform, a realistic parking model is created for data gathering. An occupancy grid network then processes this data to predict vehicle paths and obstacles, enhancing the system's perception in complex indoor environments. Ultimately, this strategy improves safety in autonomous parking operations. The paper meticulously evaluates the model's predictive capabilities, validating its efficacy in the context of underground parking. Our findings confirm that the proposed strategy successfully enhances autonomous vehicle performance in these complex indoor settings. It equips autonomous systems with improved adaptation to underground lots, reinforcing safety measures and dependability. This work paves the way for future advancements and applications by addressing the research shortfall concerning indoor parking environments, serving as a pivotal reference point.
- G. Lan and Q. Hao, “End-to-end planning of autonomous driving in industry and academia: 2022-2023,” arXiv e-prints, pp. arXiv–2401, 2023.
- Z. Liu, H. Tang, A. Amini, X. Yang, H. Mao, D. L. Rus, and S. Han, “Bevfusion: Multi-task multi-sensor fusion with unified bird’s-eye view representation,” in 2023 IEEE international conference on robotics and automation (ICRA). IEEE, 2023, pp. 2774–2781.
- G. Lan, Y. Wu, F. Hu, and Q. Hao, “Vision-based human pose estimation via deep learning: a survey,” IEEE Transactions on Human-Machine Systems, vol. 53, no. 1, pp. 253–268, 2022.
- Y. Wei, L. Zhao, W. Zheng, Z. Zhu, J. Zhou, and J. Lu, “Surroundocc: Multi-camera 3d occupancy prediction for autonomous driving,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 21 729–21 740.
- T. Liu, G. Lan, K. A. Feenstra, Z. Huang, and J. Heringa, “Towards a knowledge graph for pre-/probiotics and microbiota–gut–brain axis diseases,” Scientific Reports, vol. 12, no. 1, p. 18977, 2022.
- H. Yi, T. Liu, and G. Lan, “The key artificial intelligence technologies in early childhood education: a review,” Artificial Intelligence Review, vol. 57, no. 1, p. 12, 2024.
- J. Philion and S. Fidler, “Lift, splat, shoot: Encoding images from arbitrary camera rigs by implicitly unprojecting to 3d,” in Computer Vision–ECCV 2020: 16th European Conference, Glasgow, UK, August 23–28, 2020, Proceedings, Part XIV 16. Springer, 2020, pp. 194–210.
- Y. Li, Z. Ge, G. Yu, J. Yang, Z. Wang, Y. Shi, J. Sun, and Z. Li, “Bevdepth: Acquisition of reliable depth for multi-view 3d object detection,” in Proceedings of the AAAI Conference on Artificial Intelligence, vol. 37, no. 2, 2023, pp. 1477–1485.
- J. Huang and G. Huang, “Bevdet4d: Exploit temporal cues in multi-camera 3d object detection,” arXiv preprint arXiv:2203.17054, 2022.
- K. Han, A. Xiao, E. Wu, J. Guo, C. Xu, and Y. Wang, “Transformer in transformer,” Advances in neural information processing systems, vol. 34, pp. 15 908–15 919, 2021.
- Y. Wang, V. C. Guizilini, T. Zhang, Y. Wang, H. Zhao, and J. Solomon, “Detr3d: 3d object detection from multi-view images via 3d-to-2d queries,” in Conference on Robot Learning. PMLR, 2022, pp. 180–191.
- H. Xu, G. Lan, S. Wu, and Q. Hao, “Online intelligent calibration of cameras and lidars for autonomous driving systems,” in 2019 IEEE Intelligent Transportation Systems Conference (ITSC). IEEE, 2019, pp. 3913–3920.
- Z. Li, W. Wang, H. Li, E. Xie, C. Sima, T. Lu, Y. Qiao, and J. Dai, “Bevformer: Learning bird’s-eye-view representation from multi-camera images via spatiotemporal transformers,” in European conference on computer vision. Springer, 2022, pp. 1–18.
- L. Mescheder, M. Oechsle, M. Niemeyer, S. Nowozin, and A. Geiger, “Occupancy networks: Learning 3d reconstruction in function space,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2019, pp. 4460–4470.
- A.-Q. Cao and R. De Charette, “Monoscene: Monocular 3d semantic scene completion,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 3991–4001.
- Y. Huang, W. Zheng, Y. Zhang, J. Zhou, and J. Lu, “Tri-perspective view for vision-based 3d semantic occupancy prediction,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2023, pp. 9223–9232.
- Y. Zhang, Z. Zhu, and D. Du, “Occformer: Dual-path transformer for vision-based 3d semantic occupancy prediction,” in Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023, pp. 9433–9443.
- Y. Li, Z. Yu, C. Choy, C. Xiao, J. M. Alvarez, S. Fidler, C. Feng, and A. Anandkumar, “Voxformer: Sparse voxel transformer for camera-based 3d semantic scene completion,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2023, pp. 9087–9098.
- H. Caesar, V. Bankiti, A. H. Lang, S. Vora, V. E. Liong, Q. Xu, A. Krishnan, Y. Pan, G. Baldan, and O. Beijbom, “nuscenes: A multimodal dataset for autonomous driving,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 11 621–11 631.
- G. Lan, Q. Lai, B. Bai, Z. Zhao, and Q. Hao, “A virtual reality training system for automotive engines assembly and disassembly,” IEEE Transactions on Learning Technologies, 2023.
- G. Lan, J. Chen, and A. Eiben, “Simulated and real-world evolution of predator robots,” in 2019 IEEE Symposium Series on Computational Intelligence (SSCI). IEEE, 2019, pp. 1974–1981.
- G. Lan, T. Liu, X. Wang, X. Pan, and Z. Huang, “A semantic web technology index,” Scientific reports, vol. 12, no. 1, p. 3672, 2022.
- G. Lan, J. Benito-Picazo, D. M. Roijers, E. Domínguez, and A. Eiben, “Real-time robot vision on low-performance computing hardware,” in 2018 15th international conference on control, automation, robotics and vision (ICARCV). IEEE, 2018, pp. 1959–1965.
- G. Lan, L. De Vries, and S. Wang, “Evolving efficient deep neural networks for real-time object recognition,” in 2019 IEEE Symposium Series on Computational Intelligence (SSCI). IEEE, 2019, pp. 2571–2578.
- G. Lan, Z. Gao, L. Tong, and T. Liu, “Class binarization to neuroevolution for multiclass classification,” Neural Computing and Applications, vol. 34, no. 22, pp. 19 845–19 862, 2022.
- Z. Gao and G. Lan, “A neat-based multiclass classification method with class binarization,” in Proceedings of the genetic and evolutionary computation conference companion, 2021, pp. 277–278.
- K. He, X. Zhang, S. Ren, and J. Sun, “Deep residual learning for image recognition,” in 2016 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), Jun 2016.
- X. Zhu, W. Su, L. Lu, B. Li, X. Wang, and J. Dai, “Deformable DETR: Deformable transformers for end-to-end object detection,” arXiv: Computer Vision and Pattern Recognition,arXiv: Computer Vision and Pattern Recognition, Oct 2020.
- G. Lan, J. M. Tomczak, D. M. Roijers, and A. Eiben, “Time efficiency in optimization with a bayesian-evolutionary algorithm,” Swarm and Evolutionary Computation, vol. 69, p. 100970, 2022.
- G. Lan, M. De Carlo, F. van Diggelen, J. M. Tomczak, D. M. Roijers, and A. E. Eiben, “Learning directed locomotion in modular robots with evolvable morphologies,” Applied Soft Computing, vol. 111, p. 107688, 2021.
- G. Lan, M. van Hooft, M. De Carlo, J. M. Tomczak, and A. E. Eiben, “Learning locomotion skills in evolvable robots,” Neurocomputing, vol. 452, pp. 294–306, 2021.