Papers
Topics
Authors
Recent
Search
2000 character limit reached

Enhanced Radar Perception via Multi-Task Learning: Towards Refined Data for Sensor Fusion Applications

Published 9 Apr 2024 in cs.CV, cs.MM, eess.IV, and eess.SP | (2404.06165v1)

Abstract: Radar and camera fusion yields robustness in perception tasks by leveraging the strength of both sensors. The typical extracted radar point cloud is 2D without height information due to insufficient antennas along the elevation axis, which challenges the network performance. This work introduces a learning-based approach to infer the height of radar points associated with 3D objects. A novel robust regression loss is introduced to address the sparse target challenge. In addition, a multi-task training strategy is employed, emphasizing important features. The average radar absolute height error decreases from 1.69 to 0.25 meters compared to the state-of-the-art height extension method. The estimated target height values are used to preprocess and enrich radar data for downstream perception tasks. Integrating this refined radar information further enhances the performance of existing radar camera fusion models for object detection and depth estimation tasks.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (25)
  1. Y. Zhou, L. Liu, H. Zhao, M. López-Benítez, L. Yu, and Y. Yue, “Towards deep radar perception for autonomous driving: Datasets, methods, and challenges,” Sensors, vol. 22, no. 11, p. 4208, 2022.
  2. X. Chen, H. Ma, J. Wan, B. Li, and T. Xia, “Multi-view 3d object detection network for autonomous driving,” in Proceedings of the IEEE conference on Computer Vision and Pattern Recognition, 2017, pp. 1907–1915.
  3. M. Liang, B. Yang, Y. Chen, R. Hu, and R. Urtasun, “Multi-task multi-sensor fusion for 3d object detection,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2019, pp. 7345–7353.
  4. K. Qian, S. Zhu, X. Zhang, and L. E. Li, “Robust multimodal vehicle detection in foggy weather using complementary lidar and radar signals,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2021, pp. 444–453.
  5. C. R. Qi, W. Liu, C. Wu, H. Su, and L. J. Guibas, “Frustum pointnets for 3d object detection from rgb-d data,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2018, pp. 918–927.
  6. H. Caesar, V. Bankiti, A. H. Lang, S. Vora, V. E. Liong, Q. Xu, A. Krishnan, Y. Pan, G. Baldan, and O. Beijbom, “nuscenes: A multimodal dataset for autonomous driving,” in Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 11 621–11 631.
  7. J.-L. Déziel, P. Merriaux, F. Tremblay, D. Lessard, D. Plourde, J. Stanguennec, P. Goulet, and P. Olivier, “Pixset: An opportunity for 3d computer vision to go beyond point clouds with a full-waveform lidar dataset,” in 2021 IEEE International Intelligent Transportation Systems Conference (ITSC).   IEEE, 2021, pp. 2987–2993.
  8. F. Nobis, M. Geisslinger, M. Weber, J. Betz, and M. Lienkamp, “A deep learning-based radar and camera sensor fusion architecture for object detection,” in 2019 Sensor Data Fusion: Trends, Solutions, Applications (SDF).   IEEE, 2019, pp. 1–7.
  9. H. Sun, H. Feng, G. Stettinger, L. Servadei, and R. Wille, “Multi-task cross-modality attention-fusion for 2d object detection,” in 2023 IEEE 26th International Conference on Intelligent Transportation Systems (ITSC).   IEEE, 2023, pp. 3619–3626.
  10. A. Musralina, T. Zwick, and M. Harter, “Height estimation methods for object detection in automotive radar applications,” in Proceedings of the 8th World Congress on Electrical Engineering and Computer Systems and Sciences (EECSS’22).   Avestia Publishing, 2022, pp. EEE–101.
  11. A. Laribi, M. Hahn, J. Dickmann, and C. Waldschmidt, “A new height-estimation method using fmcw radar doppler beam sharpening,” in 2017 25th European Signal Processing Conference (EUSIPCO), 2017, pp. 1932–1396.
  12. ——, “A machine learning approach for radar based height estimation,” in 2018 21st International Conference on Intelligent Transportation Systems (ITSC), 2018, pp. 2364–2370.
  13. M. Horn, O. Schumann, M. Hahn, J. Dickmann, and K. Dietmayer, “Motion classification and height estimation of pedestrians using sparse radar data,” in 2018 Sensor Data Fusion: Trends, Solutions, Applications (SDF).   IEEE, 2018, pp. 1–6.
  14. C.-C. Lo and P. Vandewalle, “Depth estimation from monocular images and sparse radar using deep ordinal regression network,” in 2021 IEEE International Conference on Image Processing (ICIP).   IEEE, 2021, pp. 3343–3347.
  15. A. Laribi, M. Hahn, J. Dickmann, and C. Waldschmidt, “Vertical digital beamforming versus multipath height finding,” in 2017 IEEE MTT-S International Conference on Microwaves for Intelligent Mobility (ICMIM), 2017, pp. 99–102.
  16. ——, “A novel target-height estimation approach using radar-wave multipath propagation for automotive applications,” Advances in Radio Science, vol. 15, pp. 61–67, 2017.
  17. S. Kohnert, M. Vogt, and R. Stolle, “Fmcw radar height estimation of moving vehicles by analyzing multipath reflections,” in 2023 20th European Radar Conference (EuRAD).   IEEE, Sep. 2023. [Online]. Available: http://dx.doi.org/10.23919/EuRAD58043.2023.10289187
  18. A. Vijayaraghavan, A. Koppelaar, and F. Laghezza, “Short range height classification in fmcw radar,” in 2020 17th European Radar Conference (EuRAD), 2021, pp. 116–119.
  19. R. Nabati and H. Qi, “Centerfusion: Center-based radar and camera fusion for 3d object detection,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2021, pp. 1527–1536.
  20. L. Stäcker, P. Heidenreich, J. Rambach, and D. Stricker, “Fusion point pruning for optimized 2d object detection with radar-camera fusion,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2022, pp. 3087–3094.
  21. P. J. Huber, “Robust Estimation of a Location Parameter,” The Annals of Mathematical Statistics, vol. 35, no. 1, pp. 73 – 101, 1964. [Online]. Available: https://doi.org/10.1214/aoms/1177703732
  22. R. Girshick, “Fast r-cnn,” in Proceedings of the IEEE international conference on computer vision, 2015, pp. 1440–1448.
  23. S. Ren, K. He, R. Girshick, and J. Sun, “Faster r-cnn: Towards real-time object detection with region proposal networks,” Advances in neural information processing systems, vol. 28, 2015.
  24. K. Simonyan and A. Zisserman, “Very deep convolutional networks for large-scale image recognition,” arXiv preprint arXiv:1409.1556, 2014.
  25. T.-Y. Lin, P. Dollár, R. Girshick, K. He, B. Hariharan, and S. Belongie, “Feature pyramid networks for object detection,” in Proceedings of the IEEE conference on computer vision and pattern recognition, 2017, pp. 2117–2125.
Citations (4)

Summary

Paper to Video (Beta)

Whiteboard

No one has generated a whiteboard explanation for this paper yet.

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.