Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
124 tokens/sec
GPT-4o
8 tokens/sec
Gemini 2.5 Pro Pro
47 tokens/sec
o3 Pro
5 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

RaLiBEV: Radar and LiDAR BEV Fusion Learning for Anchor Box Free Object Detection Systems (2211.06108v5)

Published 11 Nov 2022 in cs.CV, cs.AI, and cs.LG

Abstract: In autonomous driving, LiDAR and radar are crucial for environmental perception. LiDAR offers precise 3D spatial sensing information but struggles in adverse weather like fog. Conversely, radar signals can penetrate rain or mist due to their specific wavelength but are prone to noise disturbances. Recent state-of-the-art works reveal that the fusion of radar and LiDAR can lead to robust detection in adverse weather. The existing works adopt convolutional neural network architecture to extract features from each sensor data, then align and aggregate the two branch features to predict object detection results. However, these methods have low accuracy of predicted bounding boxes due to a simple design of label assignment and fusion strategies. In this paper, we propose a bird's-eye view fusion learning-based anchor box-free object detection system, which fuses the feature derived from the radar range-azimuth heatmap and the LiDAR point cloud to estimate possible objects. Different label assignment strategies have been designed to facilitate the consistency between the classification of foreground or background anchor points and the corresponding bounding box regressions. Furthermore, the performance of the proposed object detector is further enhanced by employing a novel interactive transformer module. The superior performance of the methods proposed in this paper has been demonstrated using the recently published Oxford Radar RobotCar dataset. Our system's average precision significantly outperforms the state-of-the-art method by 13.1% and 19.0% at Intersection of Union (IoU) of 0.8 under 'Clear+Foggy' training conditions for 'Clear' and 'Foggy' testing, respectively.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (68)
  1. L. Chen, X. Hu, W. Tian, H. Wang, D. Cao and F. -Y. Wang, “Parallel planning: A new motion planning framework for autonomous driving,” IEEE/CAA Journal of Automatica Sinica, vol. 6, no. 1, pp. 236-246, Jan. 2019.
  2. D. Barnes, M. Gadd, P. Murcutt, P. Newman, and I. Posner, “The oxford radar robotcar dataset: A radar extension to the oxford robotcar dataset,” in Proceedings of IEEE International Conference on Robotics and Automation (ICRA), 2020, pp. 6433-6438.
  3. A. Li, S. Chen, L. Sun, N. Zheng, M. Tomizuka, and W. Zhan, “SceGene: Bio-inspired traffic scenario generation for autonomous driving testing,” IEEE Transactions on Intelligent Transportation Systems, vol. 23, no. 9, pp. 14859-14874, Sept. 2022.
  4. A. Chalvatzaras, I. Pratikakis, and A. A. Amanatiadis, “A survey on map-based localization techniques for autonomous vehicles,” IEEE Transactions on Intelligent Vehicles vol. 8, no. 2, pp. 1574-1596, Feb. 2023.
  5. D. J. Fremont, E. Kim, Y. V. Pant, S. A. Seshia, A. Acharya, X. Bruso, P. Wells, S. Lemke, Q. Lu, and S. Mehta. “Formal scenario-based testing of autonomous vehicles: From simulation to the real world,” in Proceedings of the IEEE International Conference on Intelligent Transportation Systems (ITSC), 2020, pp. 1-8.
  6. S. Chen, B. Liu, C. Feng, C. Vallespi-Gonzalez, and C. Wellington, “3D point cloud processing and learning for autonomous driving: Impacting map creation, localization, and perception,” IEEE Signal Processing Magazine, vol. 38, no. 1, pp. 68-86, Jan. 2021.
  7. T. Yin, X. Zhou, and P. Krahenbuhl, “Center-based 3D object detection and tracking,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 11784-11793.
  8. A. Lang, S. Vora, H. Caesar, L. Zhou, J. Yang, and O. Beijbom, “Pointpillars: Fast encoders for object detection from point clouds,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2019, pp. 12697-12705.
  9. J. Liu, L. Bai, Y. Xia, T. Huang, B. Zhu, and Q.-L. Han, “GNN-PMB: A simple but effective online 3D multi-object tracker without bells and whistles,” IEEE Transactions on Intelligent Vehicles, vol. 8, no. 2, pp. 1176-1189, Feb. 2023.
  10. Z. Zhang, J. Liu, Y. Xia, T. Huang, Q.-L. Han, and H. Liu, “LEGO: Learning and graph-optimized modular tracker for online multi-object tracking with point clouds,” 2023, arXiv:2308.09908.
  11. S. Liang, Z. Cao, C. Wang, and J. Yu, “Hierarchical estimation-based lidar odometry with scan-to-map matching and fixed-lag smoothing,” IEEE Transactions on Intelligent Vehicles, vol. 8, no. 2, pp. 1607-1623, Feb. 2023.
  12. H. Guo, J. Zhu, and Y. Chen, “E-LOAM: LiDAR odometry and mapping with expanded local structural information,” IEEE Transactions on Intelligent Vehicles, vol. 8, no. 2, pp. 1911-1921, Feb. 2023.
  13. L. Wang, X. Zhang, Z. Song, J. Bi, G. Zhang, H. Wei, L. Tang, L. Yang, J. Li, C. Jia, and L. Zhao, “Multi-Modal 3D object detection in autonomous driving: A survey and taxonomy,” IEEE Transactions on Intelligent Vehicles, vol. 8, no. 7, pp. 3781-3798, Jul. 2023.
  14. S. Richter, F. Bieder, S. Wirges, and C. Stiller, “A dual evidential top-view representation to model the semantic environment of automated vehicles,” IEEE Transactions on Intelligent Vehicles, doi: 10.1109/TIV.2023.3284400.
  15. S. Hasirlioglu and A. Riener, “A general approach for simulating rain effects on sensor data in real and virtual environments,” IEEE Transactions on Intelligent Vehicles, vol. 5, no. 3, pp. 426-438, Sept. 2020.
  16. M. Dreissig, D. Scheuble, F. Piewak, and J. Boedecker, “Survey on LiDAR perception in adverse weather conditions,” in Proceedings of the IEEE Intelligent Vehicles Symposium (IV), 2023, pp. 1-8.
  17. S. Sun, A. P. Petropulu, and H. V. Poor, “MIMO radar for advanced driver-assistance systems and autonomous driving: Advantages and challenges,” IEEE Signal Processing Magazine, vol. 37, no. 4, pp. 98-117, June 2020.
  18. T. Huang, J. Liu, X. Zhou, D. C. Nguyen, M. R. Azghadi, Y. Xia, Q.-L. Han, and S. Sun, “V2X cooperative perception for autonomous driving: Recent advances and challenges,” 2023, arXiv:2310.03525.
  19. J. Liu, Q. Zhao, W. Xiong, T. Huang, Q.-L. Han, and B. Zhu, “SMURF: Spatial multi-representation fusion for 3D object detection with 4D imaging radar,” IEEE Transactions on Intelligent Vehicles, Oct. 2023, doi: 10.1109/TIV.2023.3322729.
  20. L. Zheng, S. Li, B. Tan, L. Yang, S. Chen, L. Huang, J. Bai, X. Zhu, and Z. Ma, “RCFusion: Fusing 4D radar and camera with bird’s-eye view features for 3D object detection,” IEEE Transactions on Instrumentation and Measurement, vol. 72, pp. 1-14, May 2023.
  21. W. Xiong, J. Liu, T. Huang, Q.-L. Han, Y. Xia, and B. Zhu, “LXL: LiDAR excluded lean 3D object detection with 4D imaging radar and camera fusion,” IEEE Transactions on Intelligent Vehicles, Oct. 2023, doi: 10.1109/TIV.2023.3321240.
  22. J. Liu, G. Ding, Y. Xia, J. Sun, T. Huang, L. Xie, and B. Zhu, “Which framework is suitable for online 3D multi-object tracking for autonomous driving with automotive 4D imaging radar?” 2023, arXiv:2309.06036.
  23. K. Qian, S. Zhu, X. Zhang, and L. E. Li, “Robust multimodal vehicle detection in foggy weather using complementary lidar and radar signals,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 444-453.
  24. Y. Li, J. Park, M. O’ Toole, and K. Kitani, “Modality-agnostic learning for radar-lidar fusion in vehicle detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 918-927.
  25. C. Qi, H. Su, K. Mo, and L. J. Guibas, “Pointnet: Deep learning on point sets for 3d classification and segmentation,” in Proceedings of IEEE Conference Computer Vision Pattern Recognition (CVPR), 2017, pp. 652–660.
  26. C. Qi, L. Yi, H. Su, and L. J. Guibas, “Pointnet++: Deep hierarchical feature learning on point sets in a metric space,” in Proceedings of the Conference on Advances in Neural Information Processing Systems (NeurIPS), 2017, pp. 5099–5108.
  27. D. Maturana and S. Scherer, “Voxnet: A 3d convolutional neural network for real-time object recognition,” in Proceedings of IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2015, pp. 922-928.
  28. C. He, R. Li, S. Li, and L. Zhang, “Voxel set transformer: A set-to-set approach to 3d object detection from point clouds,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 8417-8427.
  29. W. Shi and R. Rajkumar, “Point-GNN: Graph neural network for 3d object detection in a point cloud,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 1711-1719.
  30. J. S. Hu, T. Kuai, and S. L. Waslander, “Point density-aware voxels for lidar 3d object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 8469-8478.
  31. A. Laddha, S. Gautam, S. Palombo, S. Pandey, and C. Vallespi-Gonzalez, “MVFuseNet: Improving end-to-end object detection and motion forecasting through multi-view fusion of lidar data,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 2865-2874.
  32. M. Kronauge and H. Rohling, “Fast two-dimensional CFAR procedure,” IEEE Transactions on Aerospace and Electronic Systems, vol. 49, no. 3, pp. 1817-1823, July 2013.
  33. S. Lim, S. Lee, and S. -C. Kim,“Clustering of detected targets using DBSCAN in automotive radar systems,” in Proceedings of the 19th International Radar Symposium (IRS), 2018, pp. 1-7.
  34. J. Liu, W. Xiong, L. Bai, Y. Xia, T. Huang, W. Ouyang, and B. Zhu, “Deep instance segmentation with automotive radar detection points,” IEEE Transactions on Intelligent Vehicles, vol. 8, no. 1, pp. 84-94, Jan. 2023.
  35. W. Xiong, J. Liu, Y. Xia, T. Huang, B. Zhu, and W. Xiang, “Contrastive learning for automotive mmWave radar detection points based instance segmentation,” in Proceedings of the IEEE International Conference on Intelligent Transportation Systems (ITSC), 2022, pp. 1255-1261.
  36. A. Danzer, T. Griebel, M. Bach, and K. Dietmayer, “2D car detection in radar data with PointNets,” in Proceedings of the IEEE International Conference on Intelligent Transportation Systems (ITSC), 2019, pp. 61-66.
  37. M. M. Dreher, E. Erçelik, T. Bänziger, and A. Knol, “Radar-based 2D car detection using deep neural networks,” in Proceedings of the IEEE International Conference on Intelligent Transportation Systems (ITSC), 2020, pp. 1-8.
  38. P. Svenningsson, F. Fioranelli, and A. Yarovoy, “Radar-PointGNN: Graph based object recognition for unstructured radar point-cloud data,” in Proceedings of the IEEE Radar Conference (RadarConf), 2021, pp. 1-6.
  39. M. Meyer, G. Kuschk, and S. Tomforde, “Graph convolutional networks for 3D object detection on radar data,” in Proceedings of IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), 2021, pp. 3060-3069.
  40. K. Patel, K. Rambach, T. Visentin, D. Rusev, M. Pfeiffer, and B. Yang, “Deep learning-based object classification on automotive radar spectra,” in Proceedings of the IEEE Radar Conference (RadarConf), 2019, pp. 1-6.
  41. A. Cozma, L. Morgan, M. Stolz, D. Stoeckel, and K. Rambach, “DeepHybrid: Deep learning on automotive radar spectra and reflections for object classification,” in Proceedings of IEEE Intelligent Transportation Systems Conference (ITSC), 2021, pp. 2682-2687.
  42. B. Major, D. Fontijne, A. Ansari, R. T. Sukhavasi, R. Gowaikar, M. Hamilton, S. Lee, S. Grzechnik, and S. Subramanian, “Vehicle detection with automotive radar using deep learning on range-azimuth-doppler tensors,” in Proceedings of IEEE/CVF International Conference on Computer Vision Workshop (ICCVW), 2019, pp. 924-932.
  43. C. Decourt, R. Vanrullen, D. Salle, and T. Oberlin, “DAROD: A deep automotive radar object detector on range-doppler maps,” in Proceedings of IEEE Intelligent Vehicles Symposium (IV), 2022, pp. 112-118.
  44. C. Decourt, R. Vanrullen, D. Salle, and T. Oberlin, “A recurrent CNN for online object detection on raw radar frames,” 2022, arXiv:2212.11172.
  45. P. Li, P. Wang, K. Berntorp, and H. Liu, “Exploiting temporal relations on radar perception for autonomous driving,” in Proceedings of IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 17071-17080.
  46. T.-Y. Huang, M.-C. Lee, C.-H. Yang, and T.-S. Lee, “YOLO-ORE: A deep learning-aided object recognition approach for radar systems,” IEEE Transactions on Vehicular Technology, vol. 72, no. 5, pp. 5715-5731, May 2023.
  47. M. Bijelic, T. Gruber, F. Mannan, F. Kraus, W. Ritter, K. Dietmayer, and F. Heide, “Seeing through fog without seeing fog: Deep multimodal sensor fusion in unseen adverse weather,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 11682-11692.
  48. L. Wang, X. Zhang, B. Xv, J. Zhang, R. Fu, X. Wang, L. Zhu, H. Ren, P. Lu, J. Li, and H. Liu, “InterFusion: Interaction-based 4D radar and liDAR fusion for 3D object detection,” in Proceedings of the IEEE/RSJ International Conference on Intelligent Robots and Systems (IROS), 2022, pp. 12247-12253.
  49. L. Wang, X. Zhang, J. Li, B. Xv, R. Fu, H. Chen, L. Yang, D. Jin, and L. Zhao, “Multi-modal and multi-scale fusion 3D object detection of 4D radar and LiDAR for autonomous driving,” IEEE Transactions on Vehicular Technology, vol. 72, no. 5, pp. 5628-5641, May 2023.
  50. A. Vaswani, N. Shazeer, N. Parmar, J. Uszkoreit, L. Jones, A. N. Gomez, Ł. Kaiser, and I. Polosukhin, “Attention is all you need,” in Proceedings of the Conference on Advances in Neural Information Processing Systems (NeurIPS), 2017, pp. 5998–6008.
  51. Y. Li, M. O’ Toole, and K. Kitani, “ST-MVDNet++: Improve vehicle detection with lidar-radar geometrical augmentation via self-training,” in Proceedings of the IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), 2023, pp. 1-5.
  52. Y. Wang, J. Deng, Y. Li, J. Hu, C. Liu, Y. Zhang, J. Ji, W. Ouyang, and Y. Zhang, “Bi-LRFusion: Bi-directional LiDAR-radar fusion for 3D dynamic object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023, pp. 13394-13403.
  53. S. Li, C. H. He, H. Chen, R. H. Li, and L. Zhang, “A dual weighting label assignment scheme for object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2022, pp. 9387-9396.
  54. R. Girshick, “Fast R-CNN,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2015, pp. 1440-1448.
  55. J. Redmon, S. Divvala, R. Girshick, and A. Farhadi, “You only look once: Unified, real-time object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2016, pp. 779-788.
  56. Z. Tian, C. Shen, H. Chen, and T. He, “FCOS: Fully convolutional one-stage object detection,” in Proceedings of the International Conference on Computer Vision (ICCV), 2019, pp. 9626-9635.
  57. K. Duan, S. Bai, L. Xie, H. Qi, Q. Huang, and Q. Tian, “CenterNet: Keypoint triplets for object detection,” in Proceedings of the International Conference on Computer Vision (ICCV), 2019, pp. 6568-6577.
  58. H. Law and J. Deng, “CornerNet: Detecting objects as paired keypoints,” in Proceedings of the European Conference on Computer Vision (ECCV), 2020, pp. 642–656.
  59. X. Zhang, F. Wan, C. Liu, R. Ji, and Q. Ye, “FreeAnchor: Learning to match anchors for visual object detection,” Advances in Neural Information Processing Systems, 2019, pp. 147-155.
  60. S. Zhang, C. Chi, Y. Yao, Z. Lei, and S. Li, “Bridging the gap between anchor-based and anchor-free detection via adaptive training sample selection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2020, pp. 9759-9768.
  61. B. Zhu, J. Wang, Z. Jiang, F. Zong, S. Liu, Z. Li, and J. Sun, “AutoAssign: Differentiable label assignment for dense object detection,” 2020, arXiv:2007.03496.
  62. C. H. Nguyen, T. C. Nguyen, T. N. Tang, and N. L. Phan, “Improving object detection by label assignment distillation,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision (WACV), 2022, pp. 1005-1014.
  63. Z. Ge, S. Liu, Z. Li, O. Yoshie, and J. Sun, “OTA: Optimal transport assignment for object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2021, pp. 303-312.
  64. A. Bochkovskiy, C. Wang, and H. Liao, “YOLOv4: Optimal speed and accuracy of object detection,” 2020, arXiv:2004.10934.
  65. T.-Y. Lin, P. Goyal, R. Girshick, K. He, and P. Dollár, “Focal loss for dense object detection,” in Proceedings of the IEEE International Conference on Computer Vision (ICCV), 2017, pp. 2980-2988.
  66. D. Erhan, C. Szegedy, A. Toshev, and D. Anguelov “Scalable object detection using deep neural networks,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2014, pp. 2147-2154.
  67. N. Carion, F. Massa, G. Synnaeve, N. Usunier, A. Kirillov, and S. Zagoruyko, “End-to-end object detection with transformers,” in Proceedings of the European Conference on Computer Vision (ECCV), 2020, pp. 213-229.
  68. J. Li, C. Luo, and X. Yang, “PillarNeXt: Rethinking network designs for 3D object detection in LiDAR point clouds,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), 2023, pp. 17567-17576.
Citations (19)

Summary

We haven't generated a summary for this paper yet.