Deep unsupervised part-whole relational visual saliency

被引:30
|
作者
Liu, Yi [1 ,2 ]
Dong, Xiaohui [1 ,2 ]
Zhang, Dingwen [3 ]
Xu, Shoukun [1 ,2 ]
机构
[1] Changzhou Univ, Aliyun Sch Big Data, Sch Comp Sci & Artificial Intelligence, Changzhou 213000, Jiangsu, Peoples R China
[2] Changzhou Univ, Sch Software, Changzhou 213000, Jiangsu, Peoples R China
[3] Northwestern Polytech Univ, Sch Automat, Xian 710129, Shaanxi, Peoples R China
基金
中国国家自然科学基金; 国家重点研发计划;
关键词
Unsupervised salient object detection; Part-object relationship; Consistency-aware fusion strategy; OBJECT DETECTION;
D O I
10.1016/j.neucom.2023.126916
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep Supervised Salient Object Detection (SSOD) excessively relies on large-scale annotated pixel-level labels which consume intensive labour acquiring high quality labels. In such precondition, deep Unsupervised Salient Object Detection (USOD) draws public attention. Under the framework of the existing deep USOD methods, they mostly generate pseudo labels by fusing several hand-crafted detectors' results. On top of that, a Fully Convolutional Network (FCN) will be trained to detect salient regions separately. While the existing USOD methods have achieved some progress, there are still challenges for them towards satisfactory performance on the complex scene, including (1) poor object wholeness owing to neglecting the hierarchy of those salient regions; (2) unsatisfactory pseudo labels causing by unprimitive fusion of hand-crafted results. To address these issues, in this paper, we introduce the property of part-whole relations endowed by a Belief Capsule Network (BCNet) for deep USOD, which is achieved by a multi-stream capsule routing strategy with a belief score for each stream within the CapsNets architecture. To train BCNet well, we generate high-quality pseudo labels from multiple hand-crafted detectors by developing a consistency-aware fusion strategy. Concretely, a weeding out criterion is first defined to filter out unreliable training samples based on the inter-method consistency among four hand-crafted saliency maps. In the following, a dynamic fusion mechanism is designed to generate high-quality pseudo labels from the remaining samples for BCNet training. Experiments on five public datasets illustrate the superiority of the proposed method. Codes have been released on: https://github.com/Mirlongue/Deep-Unsupervised-Part-Whole-Relational-Visual-Saliency.
引用
收藏
页数:13
相关论文
共 50 条
  • [1] PART-WHOLE RELATIONSHIPS IN VISUAL-SEARCH
    WOLFE, JM
    FRIEDMANHILL, SR
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 1992, 33 (04) : 1355 - 1355
  • [2] Part-Object Relational Visual Saliency
    Liu, Yi
    Zhang, Dingwen
    Zhang, Qiang
    Han, Jungong
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2022, 44 (07) : 3688 - 3704
  • [3] Part-whole science
    Winther, Rasmus Gronfeldt
    SYNTHESE, 2011, 178 (03) : 397 - 427
  • [4] A reversal of part-whole Gestalts in a patient with visual form agnosia
    de-Wit, L.
    Kubilius, J.
    Op de Beeck, H.
    Wagemans, J.
    PERCEPTION, 2012, 41 (12) : 1527 - 1527
  • [5] Part-whole science
    Rasmus Grønfeldt Winther
    Synthese, 2011, 178 : 397 - 427
  • [6] PART-WHOLE RELATIONSHIPS IN THE PROCESSING OF SMALL VISUAL-PATTERNS
    JOHNSON, NF
    TURNERLYGA, M
    PETTEGREW, BS
    MEMORY & COGNITION, 1986, 14 (01) : 5 - 16
  • [7] From sublexical features to syntax: Resumption, relational nouns and part-whole nouns
    Agullo, Jorge
    REVISTA SIGNOS, 2019, 52 (100): : 590 - 613
  • [8] Part-Whole Relational Fusion Towards Multi-Modal Scene Understanding
    Liu, Yi
    Li, Chengxin
    Xu, Shoukun
    Han, Jungong
    INTERNATIONAL JOURNAL OF COMPUTER VISION, 2025,
  • [9] TRANSITIVITY OF THE PART-WHOLE RELATION
    CRUSE, DA
    JOURNAL OF LINGUISTICS, 1979, 15 (01) : 29 - 38
  • [10] PART-WHOLE TRANSFER WITH BILINGUALS
    SAEGERT, J
    KAZARIAN, S
    YOUNG, RK
    AMERICAN JOURNAL OF PSYCHOLOGY, 1973, 86 (03): : 537 - 546