Camouflaged Object Detection Based on Deep Learning with Attention-Guided Edge Detection and Multi-Scale Context Fusion

被引:1
|
作者
Wen, Yalin [1 ]
Ke, Wei [1 ,2 ]
Sheng, Hao [1 ,3 ,4 ]
机构
[1] Macao Polytech Univ, Fac Appl Sci, Macau 999078, Peoples R China
[2] Macao Polytech Univ, Engn Res Ctr Appl Technol Machine Translat & Artif, Minist Educ, Macau 999078, Peoples R China
[3] Beihang Univ, Sch Comp Sci & Engn, State Key Lab Virtual Real Technol & Syst, Beijing 100191, Peoples R China
[4] Beihang Univ, Zhongfa Aviat Inst, Hangzhou 310000, Peoples R China
来源
APPLIED SCIENCES-BASEL | 2024年 / 14卷 / 06期
关键词
camouflaged object detection; EfficientNet; salient object detection; deep learning; NETWORK;
D O I
10.3390/app14062494
中图分类号
O6 [化学];
学科分类号
0703 ;
摘要
In nature, objects that use camouflage have features like colors and textures that closely resemble their background. This creates visual illusions that help them hide and protect themselves from predators. This similarity also makes the task of detecting camouflaged objects very challenging. Methods for camouflaged object detection (COD), which rely on deep neural networks, are increasingly gaining attention. These methods focus on improving model performance and computational efficiency by extracting edge information and using multi-layer feature fusion. Our improvement is based on researching ways to enhance efficiency in the encode-decode process. We have developed a variant model that combines Swin Transformer (Swin-T) and EfficientNet-B7. This model integrates the strengths of both Swin-T and EfficientNet-B7, and it employs an attention-guided tracking module to efficiently extract edge information and identify objects in camouflaged environments. Additionally, we have incorporated dense skip links to enhance the aggregation of deep-level feature information. A boundary-aware attention module has been incorporated into the final layer of the initial shallow information recognition phase. This module utilizes the Fourier transform to quickly relay specific edge information from the initially obtained shallow semantics to subsequent stages, thereby more effectively achieving feature recognition and edge extraction. In the latter phase, which is focused on deep semantic extraction, we employ a dense skip joint attention module to enhance the decoder's performance and efficiency, ensuring accurate capture of deep-level information, feature recognition, and edge extraction. In the later stage of deep semantic extraction, we use a dense skip joint attention module to improve the decoder's performance and efficiency in capturing precise deep information. This module efficiently identifies the specifics and edge information of undetected camouflaged objects across channels and spaces. Differing from previous methods, we introduce an adaptive pixel strength loss function for handling key captured information. Our proposed method shows strong competitive performance on three current benchmark datasets (CHAMELEON, CAMO, COD10K). Compared to 26 previously proposed methods using 4 measurement metrics, our approach exhibits favorable competitiveness.
引用
收藏
页数:15
相关论文
共 50 条
  • [31] Spatial attention-guided deformable fusion network for salient object detection
    Aiping Yang
    Yan Liu
    Simeng Cheng
    Jiale Cao
    Zhong Ji
    Yanwei Pang
    Multimedia Systems, 2023, 29 : 2563 - 2573
  • [32] Multi-scale Fusion with Context-aware Network for Object Detection
    Wang, Hanyuan
    Xu, Jie
    Li, Linke
    Tian, Ye
    Xu, Du
    Xu, Shizhong
    2018 24TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2018, : 2486 - 2491
  • [33] Dual attention guided multi-scale fusion network for RGB-D salient object detection
    Gao, Huan
    Guo, Jichang
    Wang, Yudong
    Dong, Jianan
    SIGNAL PROCESSING-IMAGE COMMUNICATION, 2023, 118
  • [34] Emphasizing Boundary-Positioning and Leveraging Multi-scale Feature Fusion for Camouflaged Object Detection
    Li, Songlin
    Li, Xiuhong
    Li, Zhe
    Li, Boyuan
    Zhou, Chenyu
    Chen, Fan
    Qiu, Tianchi
    Li, Zeyu
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT XII, 2024, 14436 : 508 - 519
  • [35] Deep Multi-Scale Feature Fusion Target Detection Algorithm Based on Deep Learning
    Liu Xin
    Chen Siyi
    Chen Xiaolong
    Du Xinhao
    LASER & OPTOELECTRONICS PROGRESS, 2021, 58 (12)
  • [36] A Camouflaged Object Detection Model Based on Deep Learning
    Wang, Yong
    Li, Ling
    Yang, Xin
    Wang, Xinxin
    Liu, Hui
    PROCEEDINGS OF 2020 IEEE INTERNATIONAL CONFERENCE ON ARTIFICIAL INTELLIGENCE AND INFORMATION SYSTEMS (ICAIIS), 2020, : 150 - 153
  • [37] AMFNet: Attention-Guided Multi-Scale Fusion Network for Bi-Temporal Change Detection in Remote Sensing Images
    Zhan, Zisen
    Ren, Hongjin
    Xia, Min
    Lin, Haifeng
    Wang, Xiaoya
    Li, Xin
    REMOTE SENSING, 2024, 16 (10)
  • [38] Attention-Guided Multi-Scale Fusion Network for Similar Objects Semantic Segmentation
    Yao, Fengqin
    Wang, Shengke
    Ding, Laihui
    Zhong, Guoqiang
    Li, Shu
    Xu, Zhiwei
    COGNITIVE COMPUTATION, 2024, 16 (01) : 366 - 376
  • [39] Deformable image registration with attention-guided fusion of multi-scale deformation fields
    Zhiquan He
    Yupeng He
    Wenming Cao
    Applied Intelligence, 2023, 53 : 2936 - 2950
  • [40] Attention-Guided Multi-Scale Fusion Network for Similar Objects Semantic Segmentation
    Fengqin Yao
    Shengke Wang
    Laihui Ding
    Guoqiang Zhong
    Shu Li
    Zhiwei Xu
    Cognitive Computation, 2024, 16 : 366 - 376