Boosting Camouflaged Object Detection with Dual-Task Interactive Transformer

被引:37
|
作者
Liu, Zhengyi [1 ]
Zhang, Zhili [1 ]
Tan, Yacheng [1 ]
Wu, Wei [1 ]
机构
[1] Anhui Univ, Sch Comp Sci & Technol, Hefei, Anhui, Peoples R China
关键词
camouflaged object detection; boundary detection; transformer; interactive; multi-task learning;
D O I
10.1109/ICPR56361.2022.9956724
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Camouflaged object detection intends to discover the concealed objects hidden in the surroundings. Existing methods follow the bio-inspired framework, which first locates the object and second refines the boundary. We argue that the discovery of camouflaged objects depends on the recurrent search for the object and the boundary. The recurrent processing makes the human tired and helpless, but it is just the advantage of the transformer with global search ability. Therefore, a dual-task interactive transformer is proposed to detect both accurate position of the camouflaged object and its detailed boundary. The boundary feature is considered as Query to improve the camouflaged object detection, and meanwhile the object feature is considered as Query to improve the boundary detection. The camouflaged object detection and the boundary detection are fully interacted by multi-head self-attention. Besides, to obtain the initial object feature and boundary feature, transformer-based backbones are adopted to extract the foreground and background. The foreground is just object, while foreground minus background is considered as boundary. Here, the boundary feature can be obtained from blurry boundary region of the foreground and background. Supervised by the object, the background and the boundary ground truth, the proposed model achieves state-of-the-art performance in public datasets. https://github.com/liuzywen/COD
引用
收藏
页码:140 / 146
页数:7
相关论文
共 50 条
  • [1] OAFormer: Occlusion Aware Transformer for Camouflaged Object Detection
    Yang, Xin
    Zhu, Hengliang
    Mao, Guojun
    Xing, Shuli
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1421 - 1426
  • [2] ForegroundNet: Domain Adaptive Transformer for Camouflaged Object Detection
    Liu, Zhouyong
    Luo, Shun
    Sun, Shilei
    Li, Chunguo
    Huang, Yongming
    Yang, Luxi
    IEEE SENSORS JOURNAL, 2024, 24 (14) : 21972 - 21986
  • [3] Dynamic interactive refinement network for camouflaged object detection
    Yaoqi Sun
    Lidong Ma
    Peiyao Shou
    Hongfa Wen
    YuHan Gao
    Yixiu Liu
    Chenggang Yan
    Haibing Yin
    Neural Computing and Applications, 2024, 36 : 3433 - 3446
  • [4] Dynamic interactive refinement network for camouflaged object detection
    Sun, Yaoqi
    Ma, Lidong
    Shou, Peiyao
    Wen, Hongfa
    Gao, Yuhan
    Liu, Yixiu
    Yan, Chenggang
    Yin, Haibing
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (07): : 3433 - 3446
  • [5] Key Object Detection: Unifying Salient and Camouflaged Object Detection Into One Task
    Yin, Pengyu
    Fu, Keren
    Zhao, Qijun
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2024, PT XII, 2025, 15042 : 536 - 550
  • [6] Uncertainty-Guided Transformer Reasoning for Camouflaged Object Detection
    Yang, Fan
    Zhai, Qiang
    Li, Xin
    Huang, Rui
    Luo, Ao
    Cheng, Hong
    Fan, Deng-Ping
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 4126 - 4135
  • [7] Interactive Transformer for Small Object Detection
    Wei, Jian
    Wang, Qinzhao
    Zhao, Zixu
    CMC-COMPUTERS MATERIALS & CONTINUA, 2023, 77 (02): : 1699 - 1717
  • [8] THE INTERACTIVE EFFECTS OF ALCOHOL AND MOOD ON DUAL-TASK PERFORMANCE
    MILES, C
    PORTER, K
    JONES, DM
    PSYCHOPHARMACOLOGY, 1986, 89 (04) : 432 - 435
  • [9] VERBAL ASYMMETRIES AND LEVELS OF PROCESSING IN AN INTERACTIVE DUAL-TASK PARADIGM
    ADAMS, BJ
    DUDA, PD
    PERCEPTUAL AND MOTOR SKILLS, 1985, 60 (02) : 371 - 383
  • [10] Camouflaged Object Detection
    Fan, Deng-Ping
    Ji, Ge-Peng
    Sun, Guolei
    Cheng, Ming-Ming
    Shen, Jianbing
    Shao, Ling
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2020, : 2774 - 2784