MOD-CL: Multi-label Object Detection with Constrained Loss (2403.07885v1)
Abstract: We introduce MOD-CL, a multi-label object detection framework that utilizes constrained loss in the training process to produce outputs that better satisfy the given requirements. In this paper, we use $\mathrm{MOD_{YOLO}}$, a multi-label object detection model built upon the state-of-the-art object detection model YOLOv8, which has been published in recent years. In Task 1, we introduce the Corrector Model and Blender Model, two new models that follow after the object detection process, aiming to generate a more constrained output. For Task 2, constrained losses have been incorporated into the $\mathrm{MOD_{YOLO}}$ architecture using Product T-Norm. The results show that these implementations are instrumental to improving the scores for both Task 1 and Task 2.
- Object Detection in 20 Years: A Survey. Proc. IEEE, 111(3):257–276, 2023.
- ROAD-R: the autonomous driving dataset with logical requirements. Mach. Learn., 112(9):3261–3291, 2023.
- YOLO by Ultralytics, January 2023.
- BoT-SORT: Robust Associations Multi-Pedestrian Tracking. arXiv preprint arXiv:2206.14651, 2022.
- Speeding Up Assumption-Based SAT. In Mikolás Janota and Inês Lynce, editors, SAT 2019, volume 11628 of Lecture Notes in Computer Science, pages 164–182. Springer, 2019.
Collections
Sign up for free to add this paper to one or more collections.
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.