Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
47 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

BCLNet: Bilateral Consensus Learning for Two-View Correspondence Pruning (2401.03459v1)

Published 7 Jan 2024 in cs.CV

Abstract: Correspondence pruning aims to establish reliable correspondences between two related images and recover relative camera motion. Existing approaches often employ a progressive strategy to handle the local and global contexts, with a prominent emphasis on transitioning from local to global, resulting in the neglect of interactions between different contexts. To tackle this issue, we propose a parallel context learning strategy that involves acquiring bilateral consensus for the two-view correspondence pruning task. In our approach, we design a distinctive self-attention block to capture global context and parallel process it with the established local context learning module, which enables us to simultaneously capture both local and global consensuses. By combining these local and global consensuses, we derive the required bilateral consensus. We also design a recalibration block, reducing the influence of erroneous consensus information and enhancing the robustness of the model. The culmination of our efforts is the Bilateral Consensus Learning Network (BCLNet), which efficiently estimates camera pose and identifies inliers (true correspondences). Extensive experiments results demonstrate that our network not only surpasses state-of-the-art methods on benchmark datasets but also showcases robust generalization abilities across various feature extraction techniques. Noteworthily, BCLNet obtains 3.98\% mAP5${\circ}$ gains over the second best method on unknown outdoor dataset, and obviously accelerates model training speed. The source code will be available at: https://github.com/guobaoxiao/BCLNet.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (31)
  1. MAGSAC: marginalizing sample consensus. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 10197–10205.
  2. Gms: Grid-based motion statistics for fast, ultra-robust feature correspondence. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 4181–4190.
  3. Neural-guided RANSAC: Learning where to sample model hypotheses. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 4322–4331.
  4. Matching with PROSAC-progressive sample consensus. In 2005 IEEE Computer Society Conference on Computer Vision and Pattern Recognition, volume 1, 220–226. IEEE.
  5. Locally optimized RANSAC. In Pattern Recognition: 25th DAGM Symposium, Magdeburg, Germany, September 10-12, 2003. Proceedings 25, 236–243. Springer.
  6. MS2DG-Net: Progressive correspondence learning via multiple sparse semantics dynamic graph. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 8973–8982.
  7. Superpoint: Self-supervised interest point detection and description. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition Workshops, 224–236.
  8. Random sample consensus: a paradigm for model fitting with applications to image analysis and automated cartography. Communications of the ACM, 24(6): 381–395.
  9. Squeeze-and-excitation networks. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 7132–7141.
  10. Image matching across wide baselines: From paper to practice. International Journal of Computer Vision, 129(2): 517–547.
  11. A survey of state-of-the-art on visual SLAM. Expert Systems with Applications, 205: 117734.
  12. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980.
  13. Progressive Neighbor Consistency Mining for Correspondence Pruning. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 9527–9537.
  14. Lowe, D. G. 2004. Distinctive image features from scale-invariant keypoints. International Journal of Computer Vision, 60: 91–110.
  15. Image matching from handcrafted to deep features: A survey. International Journal of Computer Vision, 129: 23–79.
  16. Locality preserving matching. International Journal of Computer Vision, 127: 512–531.
  17. Pytorch: An imperative style, high-performance deep learning library. Advances in Neural Information Processing Systems, 32.
  18. Pointnet++: Deep hierarchical feature learning on point sets in a metric space. Advances in Neural Information Processing Systems, 30.
  19. Deep fundamental matrix estimation. In Proceedings of the European Conference on Computer Vision, 284–299.
  20. YFCC100M: The new data in multimedia research. Communications of the ACM, 59(2): 64–73.
  21. MLESAC: A new robust estimator with application to estimating image geometry. Computer Vision and Image Understanding, 78(1): 138–156.
  22. Attention is all you need. Advances in Neural Information Processing Systems, 30.
  23. Dynamic graph cnn for learning on point clouds. ACM Transactions on Graphics (tog), 38(5): 1–12.
  24. Segmentation by continuous latent semantic analysis for multi-structure model fitting. International Journal of Computer Vision, 129(7): 2034–2056.
  25. Sun3d: A database of big spaces reconstructed using sfm and object labels. In Proceedings of the IEEE International Conference on Computer Vision, 1625–1632.
  26. A decomposition model for stereo matching. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 6091–6100.
  27. Learning to find good correspondences. In Proceedings of the IEEE Conference on Computer Vision and Pattern Recognition, 2666–2674.
  28. Learning two-view correspondences and geometry using order-aware network. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 5845–5854.
  29. ConvMatch: Rethinking Network Design for Two-View Correspondence Learning. In Proc. AAAI Conf. Artif. Intell, 1–12.
  30. Progressive correspondence pruning by consensus learning. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 6464–6473.
  31. MSA-Net: Establishing reliable correspondences by multiscale attention network. IEEE Transactions on Image Processing, 31: 4598–4608.
Citations (4)

Summary

We haven't generated a summary for this paper yet.