GA-Net: Accurate and Efficient Object Detection on UAV Images Based on Grid Activations

被引:3
|
作者
Zhang, Ruiyi [1 ]
Luo, Bin [1 ]
Su, Xin [2 ]
Liu, Jun [1 ]
机构
[1] Wuhan Univ, State Key Lab Informat Engn Surveying Mapping & Re, Wuhan 430079, Peoples R China
[2] Wuhan Univ, Sch Remote Sensing & Informat Engn, Wuhan 430079, Peoples R China
基金
中国国家自然科学基金;
关键词
drone-view object detection; real-time inference; background bias mitigation; NETWORK;
D O I
10.3390/drones8030074
中图分类号
TP7 [遥感技术];
学科分类号
081102 ; 0816 ; 081602 ; 083002 ; 1404 ;
摘要
Object detection plays a crucial role in unmanned aerial vehicle (UAV) missions, where captured objects are often small and require high-resolution processing. However, this requirement is always in conflict with limited computing resources, vast fields of view, and low latency requirements. To tackle these issues, we propose GA-Net, a novel approach tailored for UAV images. The key innovation includes the Grid Activation Module (GAM), which efficiently calculates grid activations, the probability of foreground presence at grid scale. With grid activations, the GAM helps filter out patches without objects, minimize redundant computations, and improve inference speeds. Additionally, the Grid-based Dynamic Sample Selection (GDSS) focuses the model on discriminating positive samples and hard negatives, addressing background bias during training. Further enhancements involve GhostFPN, which refines Feature Pyramid Network (FPN) using Ghost module and depth-wise separable convolution. This not only expands the receptive field for improved accuracy, but also reduces computational complexity. We conducted comprehensive evaluations on DGTA-Cattle-v2, a synthetic dataset with added background images, and three public datasets (VisDrone, SeaDronesSee, DOTA) from diverse domains. The results prove the effectiveness and practical applicability of GA-Net. Despite the common accuracy and speed trade-off challenge, our GA-Net successfully achieves a mutually beneficial scenario through the strategic use of grid activations.
引用
收藏
页数:29
相关论文
共 50 条
  • [41] Object Detection in UAV Images Based on RT-DETR with CG Downsampling and CCFMP
    Yu, Chushi
    Shin, Yoan
    2024 IEEE VTS ASIA PACIFIC WIRELESS COMMUNICATIONS SYMPOSIUM, APWCS 2024, 2024,
  • [42] Improved YOLOv7-Tiny for Object Detection Based on UAV Aerial Images
    Zhang, Zitong
    Xie, Xiaolan
    Guo, Qiang
    Xu, Jinfan
    ELECTRONICS, 2024, 13 (15)
  • [43] Multi-scale object detection in UAV images based on adaptive feature fusion
    Tan, Siqi
    Duan, Zhijian
    Pu, Longzhong
    PLOS ONE, 2024, 19 (03):
  • [44] Receptive Field Block Net for Accurate and Fast Object Detection
    Liu, Songtao
    Huang, Di
    Wang, Yunhong
    COMPUTER VISION - ECCV 2018, PT XI, 2018, 11215 : 404 - 419
  • [45] Low-Latency Aerial Images Object Detection for UAV
    Feng, Kai
    Li, Weixing
    Han, Jun
    Pan, Feng
    UNMANNED SYSTEMS, 2022, 10 (01) : 57 - 67
  • [46] APNet: Accurate Positioning Deformable Convolution for UAV Image Object Detection
    Zhang, Peiran
    Zhang, Guoxin
    Yang, Kuihe
    IEEE LATIN AMERICA TRANSACTIONS, 2024, 22 (04) : 304 - 311
  • [47] Ground object classification based on UAV multispectral images
    Wu, Mu Yao
    Cheng, Sifan
    Qin, Linlin
    Wu, Gang
    2019 CHINESE AUTOMATION CONGRESS (CAC2019), 2019, : 4629 - 4634
  • [48] Efficient Framework for Palm Tree Detection in UAV Images
    Malek, Salim
    Bazi, Yakoub
    Alajlan, Naif
    AlHichri, Haikel
    Melgani, Farid
    IEEE JOURNAL OF SELECTED TOPICS IN APPLIED EARTH OBSERVATIONS AND REMOTE SENSING, 2014, 7 (12) : 4692 - 4703
  • [49] Efficient Object Detection for High Resolution Images
    Lu, Yongxi
    Javidi, Tara
    2015 53RD ANNUAL ALLERTON CONFERENCE ON COMMUNICATION, CONTROL, AND COMPUTING (ALLERTON), 2015, : 1091 - 1098
  • [50] Efficient Selective Context Network for Accurate Object Detection
    Nie, Jing
    Pang, Yanwei
    Zhao, Shengjie
    Han, Jungong
    Li, Xuelong
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2021, 31 (09) : 3456 - 3468