HA-FPN: Hierarchical Attention Feature Pyramid Network for Object Detection

被引:3
|
作者
Dang, Jin [1 ]
Tang, Xiaofen [1 ]
Li, Shuai [1 ]
机构
[1] Ningxia Univ, Sch Informat Engn, Yinchuan 750021, Peoples R China
关键词
transformer; feature pyramid networks; object detection; attention modules;
D O I
10.3390/s23094508
中图分类号
O65 [分析化学];
学科分类号
070302 ; 081704 ;
摘要
The goals of object detection are to accurately detect and locate objects of various sizes in digital images. Multi-scale processing technology can improve the detection accuracy of the detector. Feature pyramid networks (FPNs) have been proven to be effective in extracting multi-scaled features. However, most existing object detection methods recognize objects in isolation, without considering contextual information between objects. Moreover, for the sake of computational efficiency, a significant reduction in the channel dimension may lead to the loss of semantic information. This study explores the utilization of attention mechanisms to augment the representational power and efficiency of features, ultimately improving the accuracy and efficiency of object detection. The study proposed a novel hierarchical attention feature pyramid network (HA-FPN), which comprises two key components: transformer feature pyramid networks (TFPNs) and channel attention modules (CAMs). In TFPNs, multi-scaled convolutional features are embedded as tokens and self-attention is applied to across both the intra- and inter-scales to capture contextual information between the tokens. CAMs are employed to select the channels with rich channel information to alleviate massive channel information losses. By introducing contextual information and attention mechanisms, the HA-FPN significantly improves the accuracy of bounding box detection, leading to more precise identification and localization of target objects. Extensive experiments conducted on the challenging MS COCO dataset demonstrate that the proposed HA-FPN outperforms existing multi-object detection models, while incurring minimal computational overhead.
引用
收藏
页数:18
相关论文
共 50 条
  • [21] Pyramid Feature Attention Network for Saliency detection
    Zhao, Ting
    Wu, Xiangqian
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 3080 - 3089
  • [22] SA-FPN: Scale-aware attention-guided feature pyramid network for small object detection on surface defect detection of steel strips
    Han, Lu
    Li, Nan
    Li, Jiahe
    Gao, Bingbing
    Niu, Dong
    MEASUREMENT, 2025, 249
  • [23] CB-FPN: object detection feature pyramid network based on context information and bidirectional efficient fusion
    Zhibo Liu
    Jian Cheng
    Pattern Analysis and Applications, 2023, 26 (3) : 1441 - 1452
  • [24] A2-FPN: Attention Aggregation based Feature Pyramid Network for Instance Segmentation
    Hu, Miao
    Li, Yali
    Fang, Lu
    Wang, Shengjin
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION, CVPR 2021, 2021, : 15338 - 15347
  • [25] A2-FPN: Attention Aggregation based Feature Pyramid Network for Instance Segmentation
    Hu, Miao
    Li, Yali
    Fang, Lu
    Wang, Shengjin
    Proceedings of the IEEE Computer Society Conference on Computer Vision and Pattern Recognition, 2021, : 15338 - 15347
  • [26] A2-FPN: Attention aggregation based feature pyramid network for instance segmentation
    Department of Electronic Engineering, Tsinghua University
    arXiv,
  • [27] Feature Pyramid Full Granularity Attention Network for Object Detection in Remote Sensing Imagery
    Liu, Chang
    Qi, Xiao
    Yin, Hang
    Song, Bowei
    Li, Ke
    Shen, Fei
    ADVANCED INTELLIGENT COMPUTING TECHNOLOGY AND APPLICATIONS, PT X, ICIC 2024, 2024, 14871 : 332 - 353
  • [28] A recursive attention-enhanced bidirectional feature pyramid network for small object detection
    Huanlong Zhang
    Qifan Du
    Qiye Qi
    Jie Zhang
    Fengxian Wang
    Miao Gao
    Multimedia Tools and Applications, 2023, 82 : 13999 - 14018
  • [29] Scale-Insensitive Object Detection via Attention Feature Pyramid Transformer Network
    Li, Lingling
    Zheng, Changwen
    Mao, Cunli
    Deng, Haibo
    Jin, Taisong
    NEURAL PROCESSING LETTERS, 2022, 54 (01) : 581 - 595
  • [30] Scale-Insensitive Object Detection via Attention Feature Pyramid Transformer Network
    Lingling Li
    Changwen Zheng
    Cunli Mao
    Haibo Deng
    Taisong Jin
    Neural Processing Letters, 2022, 54 : 581 - 595