Rich Embedding Features for One-Shot Semantic Segmentation

被引:22
|
作者
Zhang, Xiaolin [1 ]
Wei, Yunchao [2 ]
Li, Zhao [3 ]
Yan, Chenggang [4 ]
Yang, Yi [5 ]
机构
[1] Univ Technol Sydney, Australian Artificial Intelligence Inst, Sydney, NSW 2007, Australia
[2] Beijing Jiaotong Univ, Inst Informat Sci, Beijing 100044, Peoples R China
[3] Shandong Comp Sci Ctr, Shandong Artificial Intelligence Inst, Natl Supercomp Ctr Jinan, Jinan 250101, Peoples R China
[4] Hangzhou Dianzi Univ, Inst Informat & Control, Hangzhou 310018, Peoples R China
[5] Zhejiang Univ, Coll Comp Sci & Technol, CCAI, Hangzhou 310027, Peoples R China
关键词
Image segmentation; Semantics; Feature extraction; Task analysis; Prototypes; Support vector machines; Pulse modulation; Deep learning; few shot segmentation; object segmentation; Siamese network;
D O I
10.1109/TNNLS.2021.3081693
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
One-shot semantic segmentation poses the challenging task of segmenting object regions from unseen categories with only one annotated example as guidance. Thus, how to effectively construct robust feature representations from the guidance image is crucial to the success of one-shot semantic segmentation. To this end, we propose in this article a simple, yet effective approach named rich embedding features (REFs). Given a reference image accompanied with its annotated mask, our REF constructs rich embedding features of the support object from three perspectives: 1) global embedding to capture the general characteristics; 2) peak embedding to capture the most discriminative information; 3) adaptive embedding to capture the internal long-range dependencies. By combining these informative features, we can easily harvest sufficient and rich guidance even from a single reference image. In addition to REF, we further propose a simple depth-priority context module to obtain useful contextual cues from the query image. This successfully raises the performance of one-shot semantic segmentation to a new level. We conduct experiments on pattern analysis, statical modeling and computational learning (Pascal) visual object classes (VOC) 2012 and common object in context (COCO) to demonstrate the effectiveness of our approach.
引用
收藏
页码:6484 / 6493
页数:10
相关论文
共 50 条
  • [31] One-Shot Video Object Segmentation Using Attention Transfer
    Chanda, Omit
    Wang, Yang
    2019 IEEE 21ST INTERNATIONAL WORKSHOP ON MULTIMEDIA SIGNAL PROCESSING (MMSP 2019), 2019,
  • [32] A dual-channel network for cross-domain one-shot semantic segmentation via adversarial learning
    Yang, Yong
    Chen, Qiong
    Liu, Qingfa
    KNOWLEDGE-BASED SYSTEMS, 2023, 275
  • [33] ONE NAND ONE-SHOT
    LESERVE, B
    ELECTRONIC ENGINEER, 1971, 30 (03): : 54 - &
  • [34] One-shot holography
    Akers, Chris
    Levine, Adam
    Penington, Geoff
    Wildenhain, Elizabeth
    SCIPOST PHYSICS, 2024, 16 (06):
  • [35] One-Shot Decoupling
    Dupuis, Frederic
    Berta, Mario
    Wullschleger, Juerg
    Renner, Renato
    COMMUNICATIONS IN MATHEMATICAL PHYSICS, 2014, 328 (01) : 251 - 284
  • [36] MIGHTY ONE-SHOT
    COPP, RM
    HYDRAULICS & PNEUMATICS, 1972, 25 (10) : 94 - &
  • [37] One-shot analysis
    Etienne Cuche
    Yves Emery
    Frédéric Montfort
    Nature Photonics, 2009, 3 : 633 - 635
  • [38] Deformable One-shot Face Stylization via DINO Semantic Guidance
    Zhou, Yang
    Chen, Zichong
    Huang, Hui
    2024 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2024, 2024, : 7787 - 7796
  • [39] ONE-SHOT HARRY
    Weinman, Sarah
    NEW YORK TIMES BOOK REVIEW, 2022, 127 : 7 - 7
  • [40] Thrifty one-shot
    Schmid, E
    ELECTRONICS WORLD, 1999, 105 (1763): : 912 - 912