DAMI-YOLOv8l: A multi-scale detection framework for light-trapping insect pest monitoring

被引:0
|
作者
Chen, Xiao [1 ,2 ,3 ,4 ]
Yang, Xinting [2 ,3 ,4 ]
Hu, Huan [1 ,2 ,3 ,4 ]
Li, Tianjun [1 ,2 ,3 ,4 ]
Zhou, Zijie [2 ,3 ,4 ,5 ]
Li, Wenyong [2 ,3 ,4 ]
机构
[1] Shanghai Ocean Univ, Coll Informat, Shanghai 201306, Peoples R China
[2] Beijing Acad Agr & Forestry Sci, Res Ctr Informat Technol, Beijing 100097, Peoples R China
[3] Natl Engn Res Ctr Informat Technol Agr, Beijing 100097, Peoples R China
[4] Natl Engn Lab Agriprod Qual Traceabil, Beijing 100097, Peoples R China
[5] Jilin Agr Univ, Coll Informat Technol, Changchun 130118, Peoples R China
关键词
Pest detection; YOLOv8; Fusion features; Small objects; Multiple scale detection;
D O I
10.1016/j.ecoinf.2025.103067
中图分类号
Q14 [生态学(生物生态学)];
学科分类号
071012 ; 0713 ;
摘要
Insect pest detection plays a crucial role in agricultural production for accurate and early pest control, thus significantly reducing crop damage and increasing yields. However, currently the small size and multi-scale characteristics of insect pests pose significant challenges for accurate object detection using computer vision technology. To address this issue, we propose a novel framework called DAMI-YOLOv8l to detect pest in images collected by a light-trapping device. The DAMI-YOLOv8l model integrates three key innovations: the Depth-wise Multi-Scale Convolution (DMC) module, the Attentional Scale Sequence Fusion with a P2 detection layer (ASF-P2) neck structure, and a novel bounding box regression loss function named Minimum Point Distance inner Intersection over Union (MPDinner-IoU). The DMC module improves multi-scale feature extraction to enable the effective capture and merging of features across different detection scales while reducing network parameters. The ASF-P2 neck structure enhances the fusion of multi-scale features while preserving critical local information related to small-scale features. Additionally, the MPDinner-IoU loss function optimizes feature measurement for small insect pest datasets by introducing geometric correction capabilities. By leveraging these innovations, the results demonstrate that the proposed framework improves many metrics, such as mAP50 from 74.5 % to 78.2 %, mAP50:95 from 52.5 % to 57.3 %, and FPS from 109.89 to 121.12, compared with those of YOLOv8l model on the proposed LP24 dataset. Furthermore, we validate its robustness on two other public datasets related to small objects, Pest24 and VisDrone2019.
引用
收藏
页数:14
相关论文
共 32 条
  • [31] L-Yolov5: A multi-scale channel attention-based method for real-time safety helmet detection of electrical construction workers
    Li, Tianyang
    Xu, Hanwen
    Han, Yingnan
    Zhao, Yi
    Yan, Hongbin
    2023 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS, IJCNN, 2023,
  • [32] L2UWE: A Framework for the Efficient Enhancement of Low-Light Underwater Images Using Local Contrast and Multi-Scale Fusion
    Marques, Tunai Porto
    Albu, Alexandra Branzan
    2020 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS (CVPRW 2020), 2020, : 2286 - 2295