Papers
Topics
Authors
Recent
2000 character limit reached

Comparative Analysis of YOLOv9, YOLOv10 and RT-DETR for Real-Time Weed Detection (2412.13490v2)

Published 18 Dec 2024 in cs.CV

Abstract: This paper presents a comprehensive evaluation of state-of-the-art object detection models, including YOLOv9, YOLOv10, and RT-DETR, for the task of weed detection in smart-spraying applications focusing on three classes: Sugarbeet, Monocot, and Dicot. The performance of these models is compared based on mean Average Precision (mAP) scores and inference times on different GPU and CPU devices. We consider various model variations, such as nano, small, medium, large alongside different image resolutions (320px, 480px, 640px, 800px, 960px). The results highlight the trade-offs between inference time and detection accuracy, providing valuable insights for selecting the most suitable model for real-time weed detection. This study aims to guide the development of efficient and effective smart spraying systems, enhancing agricultural productivity through precise weed management.

Citations (1)

Summary

  • The paper demonstrates that YOLOv10 achieves 93.7% mAP50 at 960px resolution with 20.1ms inference on RTX3090, outperforming RT-DETR in speed while maintaining high accuracy.
  • It evaluates model performance across various image resolutions and GPU configurations to balance detection accuracy with real-time processing requirements.
  • The research highlights the potential for smart agriculture through targeted herbicide applications driven by rapid and precise weed detection.

Evaluation of YOLOv9, YOLOv10, and RT-DETR for Real-Time Weed Detection

The comparative study on YOLOv9, YOLOv10, and RT-DETR focuses on analyzing these object detection models in the specific context of real-time weed detection within agricultural fields. Such an application is pivotal for optimizing the precision in the use of herbicides and adhering to sustainable agriculture practices. By examining models for the detection of three specific classes—Sugarbeet, Monocot, and Dicot—this paper provides insights essential for improving smart agricultural spraying technologies.

Key Findings

The study presents several notable numerical findings concerning the performance of YOLOv9, YOLOv10, and RT-DETR models. The assessment relies on metrics such as mean Average Precision (mAP) across different Intersection over Union (IoU) thresholds, specifically mAP50 and mAP50-95, and the inference time across various GPU models—namely RTX3090, RTX4090, and RTX3080 laptop. The authors have chosen to span multiple image resolutions, from 320px to 960px, with a focus on balancing detection accuracy and inference times. YOLOv9 and YOLOv10 models performed robustly across image sizes, with YOLOv10 showing a slight advantage in terms of inference speed.

Specifically, YOLOv10 (l) achieved an mAP50 of 93.7% at 960 pixels image size with inference times of 20.1ms on the RTX3090. These results underscore the ability of the YOLOv10 model to deliver high accuracy rapidly, outperforming RT-DETR in inference speed despite comparable precision metrics. The RT-DETR model excels in terms of precision but demonstrates longer inference times, which could influence its utility in real-time scenarios.

Implications for Agricultural Technologies

The implications of these findings are significant for smart agriculture. The capability to quickly and accurately distinguish between crops and unwanted vegetation allows for more targeted herbicide applications, resulting in reduced chemical usage and enhanced crop yield. The insights around model performance, especially concerning YOLOv10, can lead to refined development strategies for real-time agricultural applications where speed is crucial.

Theoretical and Practical Contributions

Theoretically, this research furthers understanding in applying deep learning object detection models in complex environments like dynamic agricultural settings. Practically, it highlights the execution efficiency of YOLO models over transformer-based alternatives for time-sensitive operations. As the industry continues to adopt more sophisticated technologies, these findings aid in prioritizing model enhancements in line with field application requirements.

Future Developments

Look towards future research facilitating higher accuracy at reduced computational costs within such task-specific contexts. Continued exploration into augmenting model architectures like YOLOv10 with further efficiency improvements, or optimizing RT-DETR for enhanced real-time applications, will be essential. The integration of methodologies such as Slicing Aided Hyper Inference (SAHI) or model compression techniques could meaningfully bolster the viability of these models for comprehensive field deployment.

In summary, this comparative analysis provides a critical evaluation of influential object detection models in the specific area of agricultural weed detection, guiding both future research and practical system development towards more efficient and sustainable farming practices.

Whiteboard

Paper to Video (Beta)

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.