HCPM: Hierarchical Candidates Pruning for Efficient Detector-Free Matching (2403.12543v1)
Abstract: Deep learning-based image matching methods play a crucial role in computer vision, yet they often suffer from substantial computational demands. To tackle this challenge, we present HCPM, an efficient and detector-free local feature-matching method that employs hierarchical pruning to optimize the matching pipeline. In contrast to recent detector-free methods that depend on an exhaustive set of coarse-level candidates for matching, HCPM selectively concentrates on a concise subset of informative candidates, resulting in fewer computational candidates and enhanced matching efficiency. The method comprises a self-pruning stage for selecting reliable candidates and an interactive-pruning stage that identifies correlated patches at the coarse level. Our results reveal that HCPM significantly surpasses existing methods in terms of speed while maintaining high accuracy. The source code will be made available upon publication.
- Hpatches: A benchmark and evaluation of handcrafted and learned local descriptors. In CVPR, pages 5173–5182, 2017a.
- Hpatches: A benchmark and evaluation of handcrafted and learned local descriptors. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 5173–5182, 2017b.
- Longformer: The long-document transformer. arXiv preprint arXiv:2004.05150, 2020.
- Htmatch: An efficient hybrid transformer based graph neural network for local feature matching. Signal Processing, 204:108859, 2023.
- Learning to match features with seeded graph matching network. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 6301–6310, 2021.
- Aspanformer: Detector-free image matching with adaptive span transformer. In ECCV, pages 20–36, 2022a.
- Sparsevit: Revisiting activation sparsity for efficient high-resolution vision transformer. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023.
- Guide local feature matching by overlap estimation. In AAAI, 2022b.
- Rethinking attention with performers. arXiv preprint arXiv:2009.14794, 2020.
- Flashattention: Fast and memory-efficient exact attention with io-awareness. Advances in Neural Information Processing Systems, 2022.
- Superpoint: Self-supervised interest point detection and description. In CVPR, pages 224–236, 2018.
- D2-net: A trainable cnn for joint detection and description of local features. In CVPR, 2019.
- Unleashing vanilla vision transformer with masked image modeling for object detection. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 2023.
- Adaptive token sampling for efficient vision transformers. In European Conference on Computer Vision, pages 396–414. Springer, 2022.
- Deep residual learning for image recognition. In CVPR, pages 770–778, 2016.
- Adaptive assignment for geometry aware local feature matching. arXiv preprint arXiv:2207.08427, 2022.
- Categorical reparameterization with gumbel-softmax. arXiv preprint arXiv:1611.01144, 2016.
- Transformers are rnns: Fast autoregressive transformers with linear attention. In ICML, pages 5156–5165, 2020.
- Reformer: The efficient transformer. arXiv preprint arXiv:2001.04451, 2020.
- Composite slice transformer: An efficient transformer with composition of multi-scale multi-range attentions. In The Eleventh International Conference on Learning Representations, 2022.
- Megadepth: Learning single-view depth prediction from internet photos. In CVPR, pages 2041–2050, 2018a.
- Megadepth: Learning single-view depth prediction from internet photos. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, pages 2041–2050, 2018b.
- Feature pyramid networks for object detection. In CVPR, pages 2117–2125, 2017a.
- Focal loss for dense object detection. In ICCV, pages 2980–2988, 2017b.
- Lightglue: Local feature matching at light speed. arXiv preprint arXiv:2306.13643, 2023.
- David G Lowe. Distinctive image features from scale-invariant keypoints. IJCV, 60(2):91–110, 2004.
- The concrete distribution: A continuous relaxation of discrete random variables. arXiv preprint arXiv:1611.00712, 2016.
- R2d2: Repeatable and reliable detector and descriptor. In CVPR, 2019.
- Superglue: Learning feature matching with graph neural networks. In CVPR, pages 4938–4947, 2020.
- Quad-networks: unsupervised learning to rank for interest point detection. In CVPR, 2017.
- Efficient attention: Attention with linear complexities. In Proceedings of the IEEE/CVF winter conference on applications of computer vision, pages 3531–3539, 2021.
- Clustergnn: Cluster-based coarse-to-fine graph neural network for efficient feature matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12517–12526, 2022.
- Loftr: Detector-free local feature matching with transformers. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 2021.
- Dynamic token pruning in plain vision transformers for semantic segmentation. In Proceedings of the IEEE/CVF International Conference on Computer Vision, pages 777–786, 2023.
- Quadtree attention for vision transformers. arXiv preprint arXiv:2201.02767, 2022a.
- Patch slimming for efficient vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 12165–12174, 2022b.
- Attention is all you need. NeurIPS, 30, 2017.
- Matchformer: Interleaving attention in transformers for feature matching. In Proceedings of the Asian Conference on Computer Vision, pages 2746–2762, 2022.
- Linformer: Self-attention with linear complexity. arXiv preprint arXiv:2006.04768, 2020.
- Joint token pruning and squeezing towards more aggressive compression of vision transformers. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 2092–2101, 2023.
- Big bird: Transformers for longer sequences. Advances in neural information processing systems, 33:17283–17297, 2020.
- Patch2pix: Epipolar-guided pixel-level correspondences. In CVPR, pages 4669–4678, 2021.
Sponsor
Paper Prompts
Sign up for free to create and run prompts on this paper using GPT-5.
Top Community Prompts
Collections
Sign up for free to add this paper to one or more collections.