Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
102 tokens/sec
GPT-4o
59 tokens/sec
Gemini 2.5 Pro Pro
43 tokens/sec
o3 Pro
6 tokens/sec
GPT-4.1 Pro
50 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Input Augmentation with SAM: Boosting Medical Image Segmentation with Segmentation Foundation Model (2304.11332v2)

Published 22 Apr 2023 in cs.CV, cs.AI, and cs.LG

Abstract: The Segment Anything Model (SAM) is a recently developed large model for general-purpose segmentation for computer vision tasks. SAM was trained using 11 million images with over 1 billion masks and can produce segmentation results for a wide range of objects in natural scene images. SAM can be viewed as a general perception model for segmentation (partitioning images into semantically meaningful regions). Thus, how to utilize such a large foundation model for medical image segmentation is an emerging research target. This paper shows that although SAM does not immediately give high-quality segmentation for medical image data, its generated masks, features, and stability scores are useful for building and training better medical image segmentation models. In particular, we demonstrate how to use SAM to augment image input for commonly-used medical image segmentation models (e.g., U-Net). Experiments on three segmentation tasks show the effectiveness of our proposed SAMAug method. The code is available at \url{https://github.com/yizhezhang2000/SAMAug}.

Enhancing Medical Image Segmentation with the SAMAug Method

Recent developments in the field of computer vision have introduced a potential avenue for improving medical image segmentation through foundation models. In particular, the Segment Anything Model (SAM) emerges as a foundational tool capable of general-purpose segmentation across diverse visual contexts. Despite its broad application potential, direct application of SAM to medical images has shown limited success. The paper "Input Augmentation with SAM: Boosting Medical Image Segmentation with Segmentation Foundation Model" systematically explores the leverage of SAM-generated features to enhance the performance of specialized medical image segmentation models through a proposed method termed SAMAug.

The core challenge addressed in this work is the domain-specific intricacies that typical general-purpose models fail to capture, particularly in medical imaging where expert knowledge is crucial. SAM, trained with an extensive dataset—11 million images and over 1 billion masks—provides segmentation outputs, including features, stability scores, and masks, yet initial tests indicate inadequate performance on medical datasets directly.

SAMAug Methodology and Implementation

The SAMAug method focuses on augmenting input images using SAM-generated segmentation prior maps and boundary prior maps. These maps enrich the original input by embedding semantically relevant structures derived from SAM, hence altering conventional data augmentation paradigms which primarily rely on transformations such as rotations or cropping.

This augmentation is seamlessly integrated into the input channel of conventional medical image segmentation models like U-Net. SAMAug does not modify the parameters of SAM itself, ensuring that computation and memory footprints remain minimal compared to approaches involving fine-tuning or additional adaptation layers. Model training with SAMAug combines augmented and raw image inputs, optimizing the model’s performance via a loss function that weights contributions from different input types. This dual-input methodology allows segmentation models to benefit both from raw and SAM-augmented data, particularly impacting scenarios where SAM alone underperforms.

Experimental Validation

Empirical evidence of SAMAug's merits is presented through experiments across three benchmark datasets: Polyp, MoNuSeg, and GlaS. In polyp segmentation, SAMAug demonstrates a notable improvement in the Dice scores over the adept HSNet. For cell segmentation tasks on the MoNuSeg dataset, SAMAug enhances Advanced Jaccard Index (AJI) and F-score metrics across multiple neural network architectures, including U-Net and Attention UNet. Similarly, in gland segmentation on the GlaS dataset, SAMAug achieves higher F-score and Object Dice measures compared to traditional approaches.

These results suggest that SAMAug significantly enhances segmentation accuracy and outline its effectiveness in mitigating the shortcomings of SAM when applied straightforwardly to medical images. Visual cases substantiate that SAM's segmentation serves as a robust prior, catalyzing the downstream models to achieve higher fidelity in segmentation outputs.

Implications and Future Directions

By demonstrating an efficient and effective augmentation strategy, the paper establishes foundational work to integrate SAM with existing medical segmentation pipelines. The implications extend both practically—where existing models can be even more finely tuned for precise medical prediction—and theoretically—introducing a pathway to bolster the utility of foundation models for specialized tasks.

Future research directions might investigate optimizing the augmentation function for greater robustness, adapting SAMAug to handle other medical imaging modalities, and leveraging the learned representations for uncertainty estimation and other predictive applications within clinical settings. Moreover, scalability to real-time applications and integration within automated diagnostic systems would significantly broaden the utility of the SAMAug approach.

In summary, SAMAug marks a strategic enhancement in the field of medical image processing, capitalizing on SAM’s perceptual abilities and the domain-specific requirements of medical segmentation tasks. This underscores the potential foundation models have for elevating medical imaging standards when appropriately augmented with task-specific methodologies.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (32)
  1. WM-DOVA maps for accurate polyp highlighting in colonoscopy: Validation vs. saliency maps from physicians. Computerized Medical Imaging and Graphics, 43:99–111, 2015.
  2. Swin-Unet: Unet-like pure Transformer for medical image segmentation. In Computer Vision–ECCV 2022 Workshops, Part III, pages 205–218. Springer, 2023.
  3. A review of medical image data augmentation techniques for deep learning applications. Journal of Medical Imaging and Radiation Oncology, 65(5):545–563, 2021.
  4. Segment anything model (SAM) for digital pathology: Assess zero-shot segmentation on whole slide imaging. arXiv preprint arXiv:2304.04155, 2023.
  5. A new medical image enhancement algorithm using adaptive parameters. International Journal of Imaging Systems and Technology, 32(6):2198–2218, 2022.
  6. DeSAM: Decoupling Segment Anything Model for generalizable medical image segmentation. arXiv preprint arXiv:2306.00499, 2023.
  7. Segment Anything Model for medical images? arXiv preprint arXiv:2304.14660, 2023.
  8. Kvasir-SEG: A segmented polyp dataset. In MultiMedia Modeling: 26th International Conference, Part II, pages 451–462. Springer, 2020.
  9. SAM struggles in concealed scenes–empirical study on ”segment anything”. arXiv preprint arXiv:2304.06022, 2023.
  10. Adam: A method for stochastic optimization. arXiv preprint arXiv:1412.6980, 2014.
  11. Segment anything. arXiv preprint arXiv:2304.02643, 2023.
  12. A dataset and a technique for generalized nuclear segmentation for computational pathology. IEEE Transactions on Medical Imaging, 36(7):1550–1560, 2017.
  13. Jun Ma and Bo Wang. Segment anything in medical images. arXiv preprint arXiv:2304.12306, 2023.
  14. Attention U-Net: Learning where to look for the pancreas. In International Conference on Medical Imaging with Deep Learning, 2018.
  15. Robustness of SAM: Segment anything under corruptions and beyond. arXiv preprint arXiv:2306.07713, 2023.
  16. U-Net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention–MICCAI 2015: 18th International Conference, Part III, pages 234–241. Springer, 2015.
  17. MedGA: A novel evolutionary method for image enhancement in medical imaging systems. Expert Systems with Applications, 119:387–399, 2019.
  18. Toward embedded detection of polyps in WCE images for early diagnosis of colorectal cancer. International Journal of Computer Assisted Radiology and Surgery, 9:283–293, 2014.
  19. Gland segmentation in colon histology images: The GlaS challenge contest. Medical Image Analysis, 35:489–502, 2017.
  20. Automated polyp detection in colonoscopy videos using shape and context information. IEEE Transactions on Medical Imaging, 35(2):630–644, 2015.
  21. A benchmark for endoluminal scene segmentation of colonoscopy images. Journal of Healthcare Engineering, 2017, 2017.
  22. Tent: Fully test-time adaptation by entropy minimization. In International Conference on Learning Representations, 2021.
  23. DeepIGeoS: A deep interactive geodesic framework for medical image segmentation. IEEE Transactions on Pattern Analysis and Machine Intelligence, 41(7):1559–1572, 2018.
  24. A practical guide to the cancer genome atlas (TCGA). Statistical Genomics: Methods and Protocols, pages 111–141, 2016.
  25. Medical SAM adapter: Adapting Segment Anything Model for medical image segmentation. arXiv preprint arXiv:2304.12620, 2023.
  26. Understanding Segment Anything Model: SAM is biased towards texture rather than shape. 2023.
  27. A survey on Segment Anything Model (SAM): Vision foundation model meets prompt engineering. arXiv preprint arXiv:2306.06211, 2023.
  28. A comprehensive survey on Segment Anything Model for vision and beyond. arXiv preprint arXiv:2305.08196, 2023.
  29. HSNet: A hybrid semantic network for polyp segmentation. Computers in Biology and Medicine, 150:106173, 2022.
  30. How Segment Anything Model (SAM) boost medical image segmentation? arXiv preprint arXiv:2305.03678, 2023.
  31. Data augmentation using learned transformations for one-shot medical image segmentation. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, pages 8543–8553, 2019.
  32. Can SAM segment polyps? arXiv preprint arXiv:2304.07583, 2023.
User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (5)
  1. Yizhe Zhang (127 papers)
  2. Tao Zhou (398 papers)
  3. Shuo Wang (382 papers)
  4. Peixian Liang (12 papers)
  5. Danny Z. Chen (72 papers)
Citations (63)
Github Logo Streamline Icon: https://streamlinehq.com

GitHub