Automatic dense-correspondence-to-SAM prompt conversion without fine-tuning
Develop a training-free method that automatically converts dense image correspondences (for example, patch-level matches produced by self-supervised Vision Transformers) into reliable foreground and background point prompts for the Segment Anything Model (SAM), ensuring that the conversion requires no model fine-tuning.
References
However, automatically converting dense correspondences into reliable foreground/background point prompts for SAM while requiring no model fine-tuning—remains open.
— Memory-SAM: Human-Prompt-Free Tongue Segmentation via Retrieval-to-Prompt
(2510.15849 - Chae et al., 17 Oct 2025) in Section 1 (Introduction)