PA-SAM: Prompt Adapter SAM for High-Quality Image Segmentation (2401.13051v1)
Abstract: The Segment Anything Model (SAM) has exhibited outstanding performance in various image segmentation tasks. Despite being trained with over a billion masks, SAM faces challenges in mask prediction quality in numerous scenarios, especially in real-world contexts. In this paper, we introduce a novel prompt-driven adapter into SAM, namely Prompt Adapter Segment Anything Model (PA-SAM), aiming to enhance the segmentation mask quality of the original SAM. By exclusively training the prompt adapter, PA-SAM extracts detailed information from images and optimizes the mask decoder feature at both sparse and dense prompt levels, improving the segmentation performance of SAM to produce high-quality masks. Experimental results demonstrate that our PA-SAM outperforms other SAM-based methods in high-quality, zero-shot, and open-set segmentation. We're making the source code and models available at https://github.com/xzz2/pa-sam.
- “Inpaint anything: Segment anything meets image inpainting,” arXiv preprint arXiv:2304.06790, 2023.
- “How segment anything model (sam) boost medical image segmentation?,” arXiv preprint arXiv:2305.03678, 2023.
- “Baseg: Boundary aware semantic segmentation for autonomous driving,” Neural Networks, vol. 157, pp. 460–470, 2023.
- “Can sam boost video super-resolution?,” arXiv preprint arXiv:2305.06524, 2023.
- “Matte anything: Interactive natural image matting with segment anything models,” arXiv preprint arXiv:2306.04121, 2023.
- “Let segment anything help image dehaze,” arXiv preprint arXiv:2306.15870, 2023.
- “Segment anything,” arXiv preprint arXiv:2304.02643, 2023.
- “Personalize segment anything model with one shot,” arXiv preprint arXiv:2305.03048, 2023.
- “Foodsam: Any food segmentation,” IEEE Transactions on Multimedia, 2023.
- “Segment anything in high quality,” in NeurIPS, 2023.
- “Rsprompter: Learning to prompt for remote sensing instance segmentation based on visual foundation model,” arXiv preprint arXiv:2306.16269, 2023.
- “Surgicalsam: Efficient class promptable surgical instrument segmentation,” arXiv preprint arXiv:2308.08746, 2023.
- “Sam-u: Multi-box prompts triggered uncertainty estimation for reliable sam in medical image,” arXiv preprint arXiv:2307.04973, 2023.
- “Samaug: Point prompt augmentation for segment anything model,” arXiv preprint arXiv:2307.01187, 2023.
- “Parameter-efficient transfer learning for nlp,” in International Conference on Machine Learning. PMLR, 2019, pp. 2790–2799.
- “Categorical reparameterization with gumbel-softmax,” arXiv preprint arXiv:1611.01144, 2016.
- “Neural discrete representation learning,” Advances in neural information processing systems, vol. 30, 2017.
- “Parameter-efficient orthogonal finetuning via butterfly factorization,” arXiv preprint arXiv:2311.06243, 2023.
- “Highly accurate dichotomous image segmentation,” in Computer Vision – ECCV 2022, Cham, 2022, pp. 38–56, Springer Nature Switzerland.
- “Deep interactive thin object selection,” in 2021 IEEE Winter Conference on Applications of Computer Vision (WACV), Jan 2021.
- “Towards high-resolution salient object detection,” in 2019 IEEE/CVF International Conference on Computer Vision (ICCV), Oct 2019.
- “Microsoft coco: Common objects in context,” in Computer Vision–ECCV 2014: 13th European Conference, Zurich, Switzerland, September 6-12, 2014, Proceedings, Part V 13. Springer, 2014, pp. 740–755.
- “Segmentation in the wild,” https://eval.ai/web/challenges/challenge-page/1931/overview?ref=blog.roboflow.com.
- “Grounding dino: Marrying dino with grounded pre-training for open-set object detection,” arXiv preprint arXiv:2303.05499, 2023.
- “Aim: Adapting image models for efficient video action recognition,” arXiv preprint arXiv:2302.03024, 2023.
- “Dino: Detr with improved denoising anchor boxes for end-to-end object detection,” arXiv preprint arXiv:2203.03605, 2022.
- Zhaozhi Xie (4 papers)
- Bochen Guan (10 papers)
- Weihao Jiang (12 papers)
- Muyang Yi (1 paper)
- Yue Ding (49 papers)
- Hongtao Lu (76 papers)
- Lei Zhang (1689 papers)