PLoc: A New Evaluation Criterion Based on Physical Location for Autonomous Driving Datasets (2403.19893v1)
Abstract: Autonomous driving has garnered significant attention as a key research area within artificial intelligence. In the context of autonomous driving scenarios, the varying physical locations of objects correspond to different levels of danger. However, conventional evaluation criteria for automatic driving object detection often overlook the crucial aspect of an object's physical location, leading to evaluation results that may not accurately reflect the genuine threat posed by the object to the autonomous driving vehicle. To enhance the safety of autonomous driving, this paper introduces a novel evaluation criterion based on physical location information, termed PLoc. This criterion transcends the limitations of traditional criteria by acknowledging that the physical location of pedestrians in autonomous driving scenarios can provide valuable safety-related information. Furthermore, this paper presents a newly re-annotated dataset (ApolloScape-R) derived from ApolloScape. ApolloScape-R involves the relabeling of pedestrians based on the significance of their physical location. The dataset is utilized to assess the performance of various object detection models under the proposed PLoc criterion. Experimental results demonstrate that the average accuracy of all object detection models in identifying a person situated in the travel lane of an autonomous vehicle is lower than that for a person on a sidewalk. The dataset is publicly available at https://github.com/lnyrlyed/ApolloScape-R.git
- “Out-of-distribution detection for automotive perception” In 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), 2021, pp. 2938–2943 IEEE
- “Lane marking detection via deep convolutional neural network” In Neurocomputing 280 Elsevier, 2018, pp. 46–55
- “Unintended effects of autonomous driving: A study on mobility preferences in the future” In Sustainability 10.7 MDPI, 2018, pp. 2404
- “Next generation radar sensors in automotive sensor fusion systems” In 2017 Sensor Data Fusion: Trends, Solutions, Applications (SDF), 2017, pp. 1–6 IEEE
- “A review of sensor technologies for perception in automated driving” In IEEE Intelligent Transportation Systems Magazine 11.4 IEEE, 2019, pp. 94–108
- “Road curb detection and localization with monocular forward-view vehicle camera” In IEEE Transactions on Intelligent Transportation Systems 20.9 IEEE, 2018, pp. 3568–3584
- Satoshi Masuda, Hiroaki Nakamura and Kohichi Kajitani “Rule-based searching for collision test cases of autonomous vehicles simulation” In IET Intelligent Transport Systems 12.9 Wiley Online Library, 2018, pp. 1088–1095
- “Development of a self-driving car that can handle the adverse weather” In International journal of automotive technology 19 Springer, 2018, pp. 191–197
- Ayşegül Uçar, Yakup Demir and Cüneyt Güzeliş “Object recognition and detection with deep learning for autonomous driving applications” In Simulation 93.9 SAGE Publications Sage UK: London, England, 2017, pp. 759–769
- “Traffic sign recognition based on color, shape, and pictogram classification using support vector machines” In Neural Computing and Applications 30 Springer, 2018, pp. 2807–2817
- “A scenario-adaptive driving behavior prediction approach to urban autonomous driving” In Applied Sciences 7.4 MDPI, 2017, pp. 426
- “Turn signal detection during nighttime by CNN detector and perceptual hashing tracking” In IEEE Transactions on Intelligent Transportation Systems 18.12 IEEE, 2017, pp. 3303–3314
- “Moving-object detection from consecutive stereo pairs using slanted plane smoothing” In IEEE Transactions on Intelligent Transportation Systems 18.11 IEEE, 2017, pp. 3093–3102
- “Surrounding vehicle detection using an FPGA panoramic camera and deep CNNs” In IEEE Transactions on Intelligent Transportation Systems 21.12 IEEE, 2019, pp. 5110–5122
- “A sensor-fusion drivable-region and lane-detection system for autonomous vehicle navigation in challenging road scenarios” In IEEE Transactions on Vehicular Technology 63.2 IEEE, 2013, pp. 540–555
- “Rich feature hierarchies for accurate object detection and semantic segmentation” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2014, pp. 580–587
- Ross Girshick “Fast r-cnn” In Proceedings of the IEEE international conference on computer vision, 2015, pp. 1440–1448
- “Faster r-cnn: Towards real-time object detection with region proposal networks” In Advances in neural information processing systems 28, 2015
- Xudong Sun, Pengcheng Wu and Steven CH Hoi “Face detection using deep learning: An improved faster RCNN approach” In Neurocomputing 299 Elsevier, 2018, pp. 42–50
- “Revisiting rcnn: On awakening the classification power of faster rcnn” In Proceedings of the European conference on computer vision (ECCV), 2018, pp. 453–468
- “Sniffer Faster R-CNN: A Joint Camera-LiDAR Object Detection Framework with Proposal Refinement” In 2023 IEEE International Conference on Mobility, Operations, Services and Technologies (MOST), 2023, pp. 1–10 IEEE
- “Ssd: Single shot multibox detector” In Computer Vision–ECCV 2016: 14th European Conference, Amsterdam, The Netherlands, October 11–14, 2016, Proceedings, Part I 14, 2016, pp. 21–37 Springer
- “You only look once: Unified, real-time object detection” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 779–788
- “Yolov3: An incremental improvement” In arXiv preprint arXiv:1804.02767, 2018
- Alexey Bochkovskiy, Chien-Yao Wang and Hong-Yuan Mark Liao “Yolov4: Optimal speed and accuracy of object detection” In arXiv preprint arXiv:2004.10934, 2020
- Chien-Yao Wang, Alexey Bochkovskiy and Hong-Yuan Mark Liao “YOLOv7: Trainable bag-of-freebies sets new state-of-the-art for real-time object detectors” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 7464–7475
- Andreas Geiger, Philip Lenz and Raquel Urtasun “Are we ready for autonomous driving? the kitti vision benchmark suite” In 2012 IEEE conference on computer vision and pattern recognition, 2012, pp. 3354–3361 IEEE
- “The apolloscape dataset for autonomous driving” In Proceedings of the IEEE conference on computer vision and pattern recognition workshops, 2018, pp. 954–960
- “Bdd100k: A diverse driving dataset for heterogeneous multitask learning” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 2636–2645
- “SODA10M: a large-scale 2D self/semi-supervised object detection dataset for autonomous driving” In arXiv preprint arXiv:2106.11118, 2021
- “Focal loss for dense object detection” In Proceedings of the IEEE international conference on computer vision, 2017, pp. 2980–2988
- “Cascade R-CNN: High quality object detection and instance segmentation” In IEEE transactions on pattern analysis and machine intelligence 43.5 IEEE, 2019, pp. 1483–1498
- “Fcos: Fully convolutional one-stage object detection” In Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 9627–9636
- “Reppoints: Point set representation for object detection” In Proceedings of the IEEE/CVF international conference on computer vision, 2019, pp. 9657–9666
- “Bridging the gap between anchor-based and anchor-free detection via adaptive training sample selection” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2020, pp. 9759–9768
- “End-to-end object detection with transformers” In European conference on computer vision, 2020, pp. 213–229 Springer
- “Sparse r-cnn: End-to-end object detection with learnable proposals” In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021, pp. 14454–14463
- “Yolox: Exceeding yolo series in 2021” In arXiv preprint arXiv:2107.08430, 2021
- “Dense Distinct Query for End-to-End Object Detection” In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 7329–7338
- “Deep residual learning for image recognition” In Proceedings of the IEEE conference on computer vision and pattern recognition, 2016, pp. 770–778