Combating noisy labels in object detection datasets (2211.13993v3)
Abstract: The quality of training datasets for deep neural networks is a key factor contributing to the accuracy of resulting models. This effect is amplified in difficult tasks such as object detection. Dealing with errors in datasets is often limited to accepting that some fraction of examples are incorrect, estimating their confidence, and either assigning appropriate weights or ignoring uncertain ones during training. In this work, we propose a different approach. We introduce the Confident Learning for Object Detection (CLOD) algorithm for assessing the quality of each label in object detection datasets, identifying missing, spurious, mislabeled, and mislocated bounding boxes and suggesting corrections. By focusing on finding incorrect examples in the training datasets, we can eliminate them at the root. Suspicious bounding boxes can be reviewed to improve the quality of the dataset, leading to better models without further complicating their already complex architectures. The proposed method is able to point out nearly 80% of artificially disturbed bounding boxes with a false positive rate below 0.1. Cleaning the datasets by applying the most confident automatic suggestions improved mAP scores by 16% to 46%, depending on the dataset, without any modifications to the network architectures. This approach shows promising potential in rectifying state-of-the-art object detection datasets.
- A. Popowicz, K. Radlak, S. Lasota, K. Szczepankiewicz, and M. Szczepankiewicz, “Combating label noise in image data using multinet flexible confident learning,” Applied Sciences, 2022.
- J. C. Chang, S. Amershi, and E. Kamar, “Revolt: Collaborative crowdsourcing for labeling machine learning datasets,” in Proc. of the 2017 CHI Conference on Human Factors in Computing Systems, 2017, p. 2334–2346.
- O. Willers, S. Sudholt, S. Raafatnia, and S. Abrecht, “Safety concerns and mitigation approaches regarding the use of deep learning in safety-critical perception tasks,” in Computer Safety, Reliability, and Security. SAFECOMP 2020 Workshops, 2020, pp. 336–350.
- B. Frénay and A. Kabán, “A comprehensive introduction to label noise,” in The European Symposium on Artificial Neural Networks, 2014.
- I. O. for Standardization, “ISO/PAS 21448 – Road Vehicles - Safety of the intended functionality,” 2022. [Online]. Available: https://www.iso.org/standard/77490.html
- ——, “ISO/TR 4804 Road vehicles — Safety and cybersecurity for automated driving systems — design, verification and validation,” 2020. [Online]. Available: https://www.iso.org/standard/80363.html
- UL Standards & Engagement, “UL 4600, standard for evaluation of autonomous products, Edition 1.” [Online]. Available: https://ul.org/UL4600
- C. G. Northcutt, A. Athalye, and J. Mueller, “Pervasive label errors in test sets destabilize machine learning benchmarks,” ArXiv, vol. abs/2103.14749, 2021.
- C. G. Northcutt, L. Jiang, and I. L. Chuang, “Confident learning: Estimating uncertainty in dataset labels,” J. Artif. Intell. Res., vol. 70, pp. 1373–1411, 2021.
- A. Thyagarajan, E. Snorrason, C. Northcutt, and J. Mueller, “Identifying incorrect annotations in multi-label classification data,” in ICLR Workshop on Trustworthy ML, 2023.
- T. A. Lampert, A. Stumpf, and P. Gançarski, “An empirical study into annotator agreement, ground truth estimation, and algorithm evaluation,” IEEE Transactions on Image Processing, vol. 25, no. 6, pp. 2557–2572, 2016.
- S. E. Reed, H. Lee, D. Anguelov, C. Szegedy, D. Erhan, and A. Rabinovich, “Training deep neural networks on noisy labels with bootstrapping,” CoRR, vol. abs/1412.6596, 2015.
- J. Goldberger and E. Ben-Reuven, “Training deep neural-networks using a noise adaptation layer,” in ICLR, 2017.
- L. Jiang, Z. Zhou, T. Leung, L.-J. Li, and L. Fei-Fei, “Mentornet: Learning data-driven curriculum for very deep neural networks on corrupted labels,” in ICML, 2018.
- P. Chen, B. Liao, G. Chen, and S. Zhang, “Understanding and utilizing deep neural networks trained with noisy labels,” in ICML, 2019.
- B. Han, Q. Yao, X. Yu, G. Niu, M. Xu, W. Hu, I. Tsang, and M. Sugiyama, “Co-teaching: Robust training of deep neural networks with extremely noisy labels,” in NeurIPS, 2018.
- S. Chadwick and P. Newman, “Training object detectors with noisy data,” 2019 IEEE Intelligent Vehicles Symposium (IV), pp. 1319–1325, 2019.
- J. Mao, Q. Yu, Y. Yamakata, and K. Aizawa, “Noisy annotation refinement for object detection,” in BMVC, 2021.
- X. Liu, W. Li, Q. Yang, B. Li, and Y. Yuan, “Towards robust adaptive object detection under noisy annotations,” IEEE/CVF Conf. on Computer Vision and Pattern Recognition (CVPR), pp. 14 187–14 196, 2022.
- S. Wang, J. Gao, B. Li, and W. Hu, “Narrowing the gap: Improved detector training with noisy location annotations,” IEEE Transactions on Image Processing, vol. 31, pp. 6369–6380, 2022. [Online]. Available: https://doi.org/10.1109%2Ftip.2022.3211468
- U. Tkachenko, A. Thyagarajan, and J. Mueller, “Objectlab: Automated diagnosis of mislabeled images in object detection data,” in ICML Workshop on Data-centric Machine Learning, 2023.
- Samasource Impact Sourcing, “The sama-coco dataset,” https://www.sama.com/sama-coco-dataset/, 2022, accessed: 2023-04-28.
- S. Ren, K. He, R. Girshick, and J. Sun, “Faster R-CNN: Towards real-time object detection with region proposal networks,” in Proc. of the 28th International Conference on Neural Information Processing Systems - Volume 1, ser. NIPS’15, 2015, p. 91–99.
- T.-Y. Lin, P. Goyal, R. Girshick, K. He, and P. Dollár, “Focal loss for dense object detection,” in 2017 IEEE Int. Conf. on Computer Vision (ICCV), 2017, pp. 2999–3007.
- J. Redmon and A. Farhadi, “Yolov3: An incremental improvement,” 2018. [Online]. Available: https://arxiv.org/abs/1804.02767
- J. Li, C. Xiong, R. Socher, and S. Hoi, “Towards noise-resistant object detection with noisy annotations,” 2020. [Online]. Available: https://arxiv.org/abs/2003.01285
- J. Ma, Y. Ushiku, and M. Sagara, “The effect of improving annotation quality on object detection datasets: A preliminary study,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2022, pp. 4850–4859.
- M. Mathias, R. Benenson, M. Pedersoli, and L. Van Gool, “Face detection without bells and whistles,” in Computer Vision – ECCV 2014, 2014, pp. 720–735.