Papers
Topics
Authors
Recent
2000 character limit reached

YOLOv13: Real-Time Object Detection with Hypergraph-Enhanced Adaptive Visual Perception (2506.17733v1)

Published 21 Jun 2025 in cs.CV

Abstract: The YOLO series models reign supreme in real-time object detection due to their superior accuracy and computational efficiency. However, both the convolutional architectures of YOLO11 and earlier versions and the area-based self-attention mechanism introduced in YOLOv12 are limited to local information aggregation and pairwise correlation modeling, lacking the capability to capture global multi-to-multi high-order correlations, which limits detection performance in complex scenarios. In this paper, we propose YOLOv13, an accurate and lightweight object detector. To address the above-mentioned challenges, we propose a Hypergraph-based Adaptive Correlation Enhancement (HyperACE) mechanism that adaptively exploits latent high-order correlations and overcomes the limitation of previous methods that are restricted to pairwise correlation modeling based on hypergraph computation, achieving efficient global cross-location and cross-scale feature fusion and enhancement. Subsequently, we propose a Full-Pipeline Aggregation-and-Distribution (FullPAD) paradigm based on HyperACE, which effectively achieves fine-grained information flow and representation synergy within the entire network by distributing correlation-enhanced features to the full pipeline. Finally, we propose to leverage depthwise separable convolutions to replace vanilla large-kernel convolutions, and design a series of blocks that significantly reduce parameters and computational complexity without sacrificing performance. We conduct extensive experiments on the widely used MS COCO benchmark, and the experimental results demonstrate that our method achieves state-of-the-art performance with fewer parameters and FLOPs. Specifically, our YOLOv13-N improves mAP by 3.0\% over YOLO11-N and by 1.5\% over YOLOv12-N. The code and models of our YOLOv13 model are available at: https://github.com/iMoonLab/yolov13.

Summary

Overview of YOLOv13: Real-Time Object Detection with Hypergraph-Enhanced Adaptive Visual Perception

The paper presents YOLOv13, a sophisticated evolution in the field of real-time object detection, part of the YOLO (You Only Look Once) series known for its efficiency and speed. YOLOv13 introduces a significant advancement by incorporating a hypergraph-based approach to enhance global high-order correlation modeling, thus improving detection performance in complex scenarios.

The paper identifies limitations in previous YOLO models, particularly up to YOLOv12, which only model local pairwise correlations and therefore struggle in environments requiring nuanced understanding of global multi-to-multi high-order interactions. The authors tackle this issue through three primary innovations: the Hypergraph-based Adaptive Correlation Enhancement (HyperACE) mechanism, the Full-Pipeline Aggregation-and-Distribution (FullPAD) paradigm, and the redesign of feature extraction blocks using depthwise separable convolutions.

Key Contributions

  1. HyperACE Mechanism: HyperACE leverages hypergraph computation to discover latent high-order correlations across spatial positions and scales. Unlike conventional methods relying on hand-crafted hyperparameters, HyperACE adaptively constructs hyperedges, thus facilitating robust and accurate correlation modeling for dynamic and complex visual data.
  2. FullPAD Paradigm: Built on the HyperACE foundation, FullPAD ensures that correlation-enhanced features are systematically distributed through the network's architecture. This paradigm supports comprehensive feature interaction from the backbone through to the detection head, enhancing gradient propagation and thereby improving the overall detection capability.
  3. Lightweight Feature Extraction: The paper introduces depthwise separable convolution blocks as replacements for traditional large-kernel convolutions. This shift significantly reduces model parameters and FLOPs without degrading performance, enabling a more resource-efficient adaptation while maintaining high detection accuracy.

Experimental Validation and Results

The authors validate YOLOv13 using the MS COCO benchmark, a widely accepted standard in object detection evaluation. The results are compelling, showing a performance improvement with YOLOv13-N achieving 3.0% mAP gain over YOLO11-N and 1.5% over YOLOv12-N. These enhancements are realized with fewer computational resources, highlighting the model's efficiency.

The experimental setup also includes cross-domain testing using Pascal VOC 2007, demonstrating YOLOv13’s robustness and generalization capabilities. Here, YOLOv13 continues to outperform predecessors, reinforcing the effectiveness of hypergraph-enhanced strategies in diverse conditions.

Discussion and Future Implications

YOLOv13's development showcases significant progress in managing complex visual environments by extending traditional object detection frameworks with multi-to-multi high-order correlation modeling. The architecture’s adaptability and efficiency address critical real-time application demands in fields such as autonomous systems and surveillance.

Future research should explore further integration of hypergraph-based approaches with other advanced machine learning techniques, potentially marrying them with transformer architectures or large-scale LLMs for broader applications in multimodal sensing and contextual understanding. Additionally, more investigation into hardware-specific optimizations for these advanced models could unlock new capabilities in embedded and edge AI systems.

In conclusion, YOLOv13 marks a significant evolution of the YOLO series, primarily through the innovative integration of hypergraph computation for enhanced visual perception, setting a strong foundation for subsequent advancements in real-time object detection paradigms.

Whiteboard

Open Problems

We haven't generated a list of open problems mentioned in this paper yet.

Continue Learning

We haven't generated follow-up questions for this paper yet.

Collections

Sign up for free to add this paper to one or more collections.

Tweets

Sign up for free to view the 1 tweet with 0 likes about this paper.