SAM-RSP: A new few-shot segmentation method based on segment anything model and rough segmentation prompts

被引:0
|
作者
Li, Jiaguang [1 ]
Wei, Ying [1 ]
Zhang, Wei [1 ]
Shi, Zhenrui [1 ]
机构
[1] Northeastern Univ, Coll Informat Sci & Engn, Shenyang 110819, Peoples R China
关键词
Few-shot segmentation; Prompt learning; Prototype learning; Segment anything model (SAM); Semantic segmentation;
D O I
10.1016/j.imavis.2024.105214
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Few-shot segmentation (FSS) aims to segment novel classes with a few labeled images. The backbones used in existing methods are pre-trained through classification tasks on the ImageNet dataset. Although these backbones can effectively perceive the semantic categories of images, they cannot accurately perceive the regional boundaries within one image, which limits the model performance. Recently, Segment Anything Model (SAM) has achieved precise image segmentation based on point or box prompts, thanks to its excellent perception of region boundaries within one image. However, it cannot effectively provide semantic information of images. This paper proposes a new few-shot segmentation method that can effectively perceive both semantic categories and regional boundaries. This method first utilizes the SAM encoder to perceive regions and obtain the query embedding. Then the support and query images are input into a backbone pre-trained on ImageNet to perceive semantics and generate a rough segmentation prompt (RSP). This query embedding is combined with the prompt to generate a pixel-level query prototype, which can better match the query embedding. Finally, the query embedding, prompt, and prototype are combined and input into the designed multi-layer prompt transformer decoder, which is more efficient and lightweight, and can provide a more accurate segmentation result. In addition, other methods can be easily combined with our framework to improve their performance. Plenty of experiments on PASCAL-5i and COCO-20i under 1-shot and 5-shot settings prove the effectiveness of our method. Our method also achieves new state-of-the-art. Codes are available at https://github.com/Jiaguang-NE U/SAM-RSP.
引用
收藏
页数:12
相关论文
共 50 条
  • [1] Segment anything model for few-shot medical image segmentation with domain tuning
    Shi, Weili
    Zhang, Penglong
    Li, Yuqin
    Jiang, Zhengang
    COMPLEX & INTELLIGENT SYSTEMS, 2025, 11 (01)
  • [2] Learning few-shot semantic segmentation with error-filtered segment anything model
    Feng, Chen-Bin
    Lai, Qi
    Liu, Kangdao
    Su, Houcheng
    Chen, Hao
    Luo, Kaixi
    Vong, Chi-Man
    VISUAL COMPUTER, 2025,
  • [3] SnapSeg: Training-Free Few-Shot Medical Image Segmentation with Segment Anything Model
    Yu, Nanxi
    Cai, Zhiyuan
    Huang, Yijin
    Tang, Xiaoying
    TRUSTWORTHY ARTIFICIAL INTELLIGENCE FOR HEALTHCARE, TAI4H 2024, 2024, 14812 : 109 - 122
  • [4] AGSAM: Agent-Guided Segment Anything Model for Automatic Segmentation in Few-Shot Scenarios
    Zhou, Hao
    He, Yao
    Cui, Xiaoxiao
    Xie, Zhi
    BIOENGINEERING-BASEL, 2024, 11 (05):
  • [5] Prototypical Metric Segment Anything Model for Data-Free Few-Shot Semantic Segmentation
    Jiang, Zhiyu
    Yuan, Ye
    Yuan, Yuan
    IEEE SIGNAL PROCESSING LETTERS, 2024, 31 : 2800 - 2804
  • [6] Learning What Not to Segment: A New Perspective on Few-Shot Segmentation
    Lang, Chunbo
    Cheng, Gong
    Tu, Binfei
    Han, Junwei
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2022, : 8047 - 8057
  • [7] CAT-SAM: Conditional Tuning for Few-Shot Adaptation of Segment Anything Model
    Xiao, Aoran
    Xuan, Weihao
    Qi, Heli
    Xing, Yun
    Ren, Ruijie
    Zhang, Xiaoqin
    Shao, Ling
    Lu, Shijian
    COMPUTER VISION - ECCV 2024, PT XL, 2025, 15098 : 189 - 206
  • [8] SEMPNet: enhancing few-shot remote sensing image semantic segmentation through the integration of the segment anything model
    Ao, Wei
    Zheng, Shunyi
    Meng, Yan
    GISCIENCE & REMOTE SENSING, 2024, 61 (01)
  • [9] BiASAM: Bidirectional-Attention Guided Segment Anything Model for Very Few-Shot Medical Image Segmentation
    Zhou, Wei
    Guan, Guilin
    Cui, Wei
    Yi, Yugen
    IEEE SIGNAL PROCESSING LETTERS, 2025, 32 : 246 - 250
  • [10] Zero-Shot Segmentation of Eye Features Using the Segment Anything Model (SAM)
    Maquiling, Virmarie
    Byrne, Sean Anthony
    Niehorster, Diederick C.
    Nystrom, Marcus
    Kasneci, Enkelejda
    PROCEEDINGS OF THE ACM ON COMPUTER GRAPHICS AND INTERACTIVE TECHNIQUES, 2024, 7 (02)