Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
129 tokens/sec
GPT-4o
28 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
38 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Adapting Segment Anything Model for Change Detection in HR Remote Sensing Images (2309.01429v4)

Published 4 Sep 2023 in cs.CV

Abstract: Vision Foundation Models (VFMs) such as the Segment Anything Model (SAM) allow zero-shot or interactive segmentation of visual contents, thus they are quickly applied in a variety of visual scenes. However, their direct use in many Remote Sensing (RS) applications is often unsatisfactory due to the special imaging characteristics of RS images. In this work, we aim to utilize the strong visual recognition capabilities of VFMs to improve the change detection of high-resolution Remote Sensing Images (RSIs). We employ the visual encoder of FastSAM, an efficient variant of the SAM, to extract visual representations in RS scenes. To adapt FastSAM to focus on some specific ground objects in the RS scenes, we propose a convolutional adaptor to aggregate the task-oriented change information. Moreover, to utilize the semantic representations that are inherent to SAM features, we introduce a task-agnostic semantic learning branch to model the semantic latent in bi-temporal RSIs. The resulting method, SAMCD, obtains superior accuracy compared to the SOTA methods and exhibits a sample-efficient learning ability that is comparable to semi-supervised CD methods. To the best of our knowledge, this is the first work that adapts VFMs for the CD of HR RSIs.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (46)
  1. A. Kirillov, E. Mintun, N. Ravi, H. Mao, C. Rolland, L. Gustafson, T. Xiao, S. Whitehead, A. C. Berg, W.-Y. Lo et al., “Segment anything,” arXiv preprint arXiv:2304.02643, 2023.
  2. X. Zhao, W. Ding, Y. An, Y. Du, T. Yu, M. Li, M. Tang, and J. Wang, “Fast segment anything,” arXiv preprint arXiv:2306.12156, 2023.
  3. C. Zhang, D. Han, Y. Qiao, J. U. Kim, S.-H. Bae, S. Lee, and C. S. Hong, “Faster segment anything: Towards lightweight sam for mobile applications,” arXiv preprint arXiv:2306.14289, 2023.
  4. W. Ji, J. Li, Q. Bi, W. Li, and L. Cheng, “Segment anything is not always perfect: An investigation of sam on different real-world applications,” arXiv preprint arXiv:2304.05750, 2023.
  5. A. Dosovitskiy, L. Beyer, A. Kolesnikov, D. Weissenborn, X. Zhai, T. Unterthiner, M. Dehghani, M. Minderer, G. Heigold, S. Gelly et al., “An image is worth 16x16 words: Transformers for image recognition at scale,” arXiv preprint arXiv:2010.11929, 2020.
  6. D. Wen, X. Huang, F. Bovolo, J. Li, X. Ke, A. Zhang, and J. A. Benediktsson, “Change detection from very-high-spatial-resolution optical remote sensing images: Methods, applications, and future directions,” IEEE Geoscience and Remote Sensing Magazine, vol. 9, no. 4, pp. 68–101, 2021.
  7. D. Peng, Y. Zhang, and H. Guan, “End-to-end change detection for high resolution satellite images using improved unet++,” Remote Sensing, vol. 11, no. 11, p. 1382, 2019.
  8. R. C. Daudt, B. Le Saux, and A. Boulch, “Fully convolutional siamese networks for change detection,” in 2018 25th IEEE International Conference on Image Processing (ICIP).   IEEE, 2018, pp. 4063–4067.
  9. M. Zhang and W. Shi, “A feature difference convolutional neural network-based change detection method,” IEEE Transactions on Geoscience and Remote Sensing, vol. 58, no. 10, pp. 7232–7246, 2020.
  10. X. Hou, Y. Bai, Y. Li, C. Shang, and Q. Shen, “High-resolution triplet network with dynamic multiscale feature for change detection on satellite images,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 177, pp. 103–115, 2021.
  11. G. Cheng, Y. Huang, X. Li, S. Lyu, Z. Xu, Q. Zhao, and S. Xiang, “Change detection methods for remote sensing in the last decade: A comprehensive review,” arXiv preprint arXiv:2305.05813, 2023.
  12. Z. Li, C. Tang, L. Wang, and A. Y. Zomaya, “Remote sensing change detection via temporal feature interaction and guided refinement,” IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–11, 2022.
  13. D. Peng, L. Bruzzone, Y. Zhang, H. Guan, and P. He, “Scdnet: A novel convolutional network for semantic change detection in high resolution optical remote sensing imagery,” International Journal of Applied Earth Observation and Geoinformation, vol. 103, p. 102465, 2021.
  14. J. Chen, Z. Yuan, J. Peng, L. Chen, H. Huang, J. Zhu, Y. Liu, and H. Li, “Dasnet: Dual attentive fully convolutional siamese networks for change detection in high-resolution satellite images,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, vol. 14, pp. 1194–1206, 2020.
  15. Q. Shi, M. Liu, S. Li, X. Liu, F. Wang, and L. Zhang, “A deeply supervised attention metric-based network and an open aerial image dataset for remote sensing change detection,” IEEE transactions on geoscience and remote sensing, vol. 60, pp. 1–16, 2021.
  16. H. Chen, J. Song, C. Wu, B. Du, and N. Yokoya, “Exchange means change: An unsupervised single-temporal change detection framework based on intra-and inter-image patch exchange,” ISPRS Journal of Photogrammetry and Remote Sensing, vol. 206, pp. 87–105, 2023.
  17. M. Seo, H. Lee, Y. Jeon, and J. Seo, “Self-pair: Synthesizing changes from single source for object change detection in remote sensing imagery,” in Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 2023, pp. 6374–6383.
  18. L. Mou, L. Bruzzone, and X. X. Zhu, “Learning spectral-spatial-temporal features via a recurrent convolutional neural network for change detection in multispectral imagery,” IEEE Transactions on Geoscience and Remote Sensing, vol. 57, no. 2, pp. 924–935, 2018.
  19. H. Chen, C. Wu, B. Du, L. Zhang, and L. Wang, “Change detection in multisource vhr images via deep siamese convolutional multiple-layers recurrent neural network,” IEEE Transactions on Geoscience and Remote Sensing, vol. 58, no. 4, pp. 2848–2864, 2019.
  20. L. Ding, H. Guo, S. Liu, L. Mou, J. Zhang, and L. Bruzzone, “Bi-temporal semantic reasoning for the semantic change detection in hr remote sensing images,” IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–14, 2022.
  21. J. Wu, B. Li, Y. Qin, W. Ni, H. Zhang, R. Fu, and Y. Sun, “A multiscale graph convolutional network for change detection in homogeneous and heterogeneous remote sensing images,” International Journal of Applied Earth Observation and Geoinformation, vol. 105, p. 102615, 2021.
  22. X. Tang, H. Zhang, L. Mou, F. Liu, X. Zhang, X. X. Zhu, and L. Jiao, “An unsupervised remote sensing change detection method based on multiscale graph convolutional network and metric learning,” IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–15, 2021.
  23. P. Yuan, Q. Zhao, X. Zhao, X. Wang, X. Long, and Y. Zheng, “A transformer-based siamese network and an open optical dataset for semantic change detection of remote sensing images,” International Journal of Digital Earth, vol. 15, no. 1, pp. 1506–1525, 2022.
  24. J. Ma, J. Duan, X. Tang, X. Zhang, and L. Jiao, “Eatder: Edge-assisted adaptive transformer detector for remote sensing change detection,” IEEE Transactions on Geoscience and Remote Sensing, 2023.
  25. H. Chen, Z. Qi, and Z. Shi, “Remote sensing image change detection with transformers,” IEEE Transactions on Geoscience and Remote Sensing, vol. 60, pp. 1–14, 2021.
  26. K. Zhang, X. Zhao, F. Zhang, L. Ding, J. Sun, and L. Bruzzone, “Relation changes matter: Cross-temporal difference transformer for change detection in remote sensing images,” IEEE Transactions on Geoscience and Remote Sensing, 2023.
  27. L. Ding, J. Zhang, K. Zhang, H. Guo, B. Liu, and L. Bruzzone, “Joint spatio-temporal modeling for semantic change detection in remote sensing images,” arXiv preprint arXiv:2212.05245, 2022.
  28. A. Radford, J. W. Kim, C. Hallacy, A. Ramesh, G. Goh, S. Agarwal, G. Sastry, A. Askell, P. Mishkin, J. Clark et al., “Learning transferable visual models from natural language supervision,” in International conference on machine learning.   PMLR, 2021, pp. 8748–8763.
  29. W. G. C. Bandara, N. G. Nair, and V. M. Patel, “Ddpm-cd: Remote sensing change detection using denoising diffusion probabilistic models,” arXiv preprint arXiv:2206.11892, 2022.
  30. X. Wang, X. Zhang, Y. Cao, W. Wang, C. Shen, and T. Huang, “Seggpt: Segmenting everything in context,” arXiv preprint arXiv:2304.03284, 2023.
  31. X. Zou, J. Yang, H. Zhang, F. Li, L. Li, J. Gao, and Y. J. Lee, “Segment everything everywhere all at once,” arXiv preprint arXiv:2304.06718, 2023.
  32. L. P. Osco, Q. Wu, E. L. de Lemos, W. N. Gonçalves, A. P. M. Ramos, J. Li, and J. M. Junior, “The segment anything model (sam) for remote sensing applications: From zero to one shot,” International Journal of Applied Earth Observation and Geoinformation, vol. 124, p. 103540, 2023.
  33. J. Wu, R. Fu, H. Fang, Y. Liu, Z. Wang, Y. Xu, Y. Jin, and T. Arbel, “Medical sam adapter: Adapting segment anything model for medical image segmentation,” arXiv preprint arXiv:2304.12620, 2023.
  34. S. Ruder, “An overview of multi-task learning in deep neural networks,” arXiv preprint arXiv:1706.05098, 2017.
  35. L. Ding and L. Bruzzone, “Diresnet: Direction-aware residual network for road extraction in vhr remote sensing images,” IEEE Transactions on Geoscience and Remote Sensing, 2020.
  36. C. Wu, B. Du, X. Cui, and L. Zhang, “A post-classification change detection method based on iterative slow feature analysis and bayesian soft fusion,” Remote Sensing of Environment, vol. 199, pp. 241–255, 2017.
  37. H. Chen and Z. Shi, “A spatial-temporal attention-based method and a new dataset for remote sensing image change detection,” Remote Sensing, vol. 12, no. 10, 2020.
  38. S. Ji, S. Wei, and M. Lu, “Fully convolutional networks for multisource building extraction from an open aerial and satellite imagery data set,” IEEE Transactions on geoscience and remote sensing, vol. 57, no. 1, pp. 574–586, 2018.
  39. L. Shen, Y. Lu, H. Chen, H. Wei, D. Xie, J. Yue, R. Chen, S. Lv, and B. Jiang, “S2looking: A satellite side-looking dataset for building change detection,” Remote Sensing, vol. 13, no. 24, p. 5094, 2021.
  40. S. Fang, K. Li, J. Shao, and Z. Li, “Snunet-cd: A densely connected siamese network for change detection of vhr images,” IEEE Geoscience and Remote Sensing Letters, vol. 19, pp. 1–5, 2021.
  41. W. G. C. Bandara and V. M. Patel, “A transformer-based siamese network for change detection,” pp. 207–210, 2022.
  42. C. Han, C. Wu, H. Guo, M. Hu, J. Li, and H. Chen, “Change guiding network: Incorporating change prior to guide change detection in remote sensing imagery,” IEEE Journal of Selected Topics in Applied Earth Observations and Remote Sensing, 2023.
  43. S. Mittal, M. Tatarchenko, and T. Brox, “Semi-supervised semantic segmentation with high-and low-level consistency,” IEEE transactions on pattern analysis and machine intelligence, vol. 43, no. 4, pp. 1369–1379, 2019.
  44. D. Peng, L. Bruzzone, Y. Zhang, H. Guan, H. Ding, and X. Huang, “Semicdnet: A semisupervised convolutional neural network for change detection in high resolution remote-sensing images,” IEEE Transactions on Geoscience and Remote Sensing, vol. 59, no. 7, pp. 5891–5906, 2020.
  45. W. G. C. Bandara and V. M. Patel, “Revisiting consistency regularization for semi-supervised change detection in remote sensing images,” arXiv preprint arXiv:2204.08454, 2022.
  46. L. Yang, L. Qi, L. Feng, W. Zhang, and Y. Shi, “Revisiting weak-to-strong consistency in semi-supervised semantic segmentation,” in Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 2023, pp. 7236–7246.
Citations (49)

Summary

  • The paper introduces SAM-CD, a framework that adapts FastSAM with a convolutional adaptor to enhance change detection accuracy in VHR remote sensing imagery.
  • It incorporates a semantic learning branch to discern true semantic changes from natural scene variations using bi-temporal image analysis.
  • Experiments demonstrate that SAM-CD outperforms traditional fully-supervised methods in precision, recall, and F1 scores while reducing the need for extensive annotations.

Adapting Vision Foundation Models for Enhanced Change Detection in Remote Sensing

The paper "Adapting Segment Anything Model for Change Detection in VHR Remote Sensing Images" presents a novel approach for improving change detection (CD) in very high-resolution (VHR) remote sensing images. The researchers leverage the capabilities of Vision Foundation Models (VFMs), particularly the Segment Anything Model (SAM) and its variant, FastSAM, to improve the accuracy and efficiency of CD tasks in remote sensing, a domain characterized by unique imaging properties that challenge traditional vision models trained on natural scenes.

Context and Motivation

CD is a crucial task in remote sensing that involves identifying and segmenting changes between multi-temporal images of the same geographic area. This capability is instrumental in environmental monitoring, urban development, and disaster management. Traditional methods often rely on aligning changes observed across images using Convolutional Neural Networks (CNNs) and, more recently, Vision Transformers (ViTs). However, these methods typically require substantial annotated data and struggle with semantic distinctions in change detection, often conflating temporal differences with semantic alterations.

Methodology

The authors propose the SAM-CD, a change detection framework that adapts FastSAM for remote sensing applications. FastSAM is employed for its robust visual representation capabilities, capturing the semantics of ground objects in remote sensing images through a convolutional adaptor tailored for task-oriented learning. The distinct approach involves:

  1. FastSAM Adaptation: Implementing a convolutional adaptor to refine the FastSAM-derived features for more accurate detection of semantically relevant changes. The multi-scale features extracted by FastSAM are aggregated and processed through this adaptor to align them with the task objectives, overcoming the inductive biases of models trained primarily on natural images.
  2. Semantic Learning Branch: A task-agnostic semantic learning branch is introduced, shifting the CD model focus from purely detecting image differences to understanding the semantic latent within multi-temporal remote sensing images. This branch exploits temporal constraints, leveraging bi-temporal image semantics to more effectively discern between natural scene variations and true semantic changes.

Results and Implications

The SAM-CD model achieves superior performance over state-of-the-art fully-supervised methods and demonstrates a sample-efficient learning process akin to semi-supervised methods. Experimental validation across multiple datasets shows significant increases in accuracy metrics such as precision, recall, and F1 scores, underscoring the effectiveness of integrating VFMs with dedicated adaptors for domain-specific applications.

The findings suggest significant implications for the field of remote sensing and image analysis. By reducing the dependency on extensive labeled datasets, SAM-CD opens pathways for more efficient and practical applications in scenarios where large-scale annotations are impractical or impossible. Furthermore, the approach highlights the potential for VFMs to be fine-tuned or adapted to niche domains beyond their original training scope, providing a flexible foundation for future advancements.

Future Directions

Future work could explore additional VFMs, further refine the interplay between semantic and change features, and extend the adaptation strategies to other challenging domains such as medical imaging. With continued advancement, the role of VFMs in semantically rich, change-detection tasks will likely expand, potentially reshaping methodologies in fields that require precise temporal-spatial analysis. The integration of VFMs with domain-specific learning branches remains a promising direction for ongoing research and development.

Github Logo Streamline Icon: https://streamlinehq.com