Method for degraded grassland gap localization based on super-resolution reconstruction and Transformer

被引:0
|
作者
Lu J. [1 ,2 ,3 ]
Chang H. [1 ]
Lan Y. [1 ,2 ,3 ]
Wang L. [1 ]
Luo H. [1 ]
Huang J. [1 ]
Yuan J. [1 ]
机构
[1] College of Electronic Engineering, College of Artificial Intelligence, South China Agricultural University, Guangzhou
[2] Heyuan Branch, Guangdong Laboratory for Lingnan Modern Agriculture, Heyuan
[3] National Center for International Collaboration Research on Precision Agricultural Aviation Pesticides Spraying Technology (NPAAC), Guangzhou
关键词
degraded grassland; gap; super-resolution reconstruction; transformer; UAV;
D O I
10.11975/j.issn.1002-6819.202401028
中图分类号
学科分类号
摘要
Grassland restoration has been a critical initiative of ecological engineering in China. Drones have effectively facilitated the reseeding in large-scale and diversified terrains at present. However, the accuracy of gap localization has been confined to the high workload during drone operations. This study aims to propose a novel approach for gap localization of degraded grassland, termed YOLOFG (YOLO for Gap), using super-resolution reconstruction and Transformer technology. Firstly, the challenges were addressed on the limited quantity and widespread distribution of gap samples in grassland scenes, in terms of sample acquisition. Drones were also employed to combine the fieldwork and simulation. A total of 132 and 282 aerial images were collected to constitute a mixed gap dataset. Secondly, three data augmentations were selected: random angular rotation, exposure adjustment, and brightness modulation, in order to enrich the mixed gap datasets. The objective of these techniques was to enhance the generalization of the model. Furthermore, 1 242 enhanced gap images were annotated to utilize the open-source labeling tool LabelImg. The annotated data was stored, according to the format specifications of the Pascal VOC dataset. Subsequently, the blurred texture was targeted in the unmanned aerial vehicle (UAV) aerial photography dataset. The modifications were made to the YOLOv5s model, thereby completing the training of the data. Finally, the spot anchor frame information was utilized to obtain from the recognition model. The imaging principles of the gimbal camera were used to obtain the conversion relationships among multiple coordinate systems. The real-time attitude information was selected from the UAV. A geodetic coordinate resolution model was established for the aerial photography targets, thus enabling precise localization of the target gaps. According to the theoretical framework of YOLOv5s target detection, a cascaded module was introduced to extract the feature texture in the neck network. The high-resolution feature maps were generated with enhanced textures, thereby enabling the model to focus on intricate feature textures. There was a great decrease in the significant scale variations and blurred textures in the drone-captured gap images. Furthermore, ShuffleNetV2 was selected as the backbone network for YOLOv5s. The number of model parameters was reduced to integrate a highly interactive Transformer self-attention structure into the backbone network. More comprehensive mid-to-high-level visual semantic information was obtained from the drone imagery. The information interaction was also enhanced among pixel blocks. More differentiated features were ultimately extracted among pixels to enhance the capture precision for the gap edge pixels. The experimental results demonstrated that the YOLOFG model achieved a mean Average Precision (mAP) of 96.57%, indicating an improvement of 11.86, 9.65, and 6.82 percentage points, respectively, compared with the YOLOv4, YOLOv7, and YOLOv8 models. The parameter count was approximately 6.24 M, with a model size of 13.9 MB. Compared with the YOLOv5s, the model size remained comparable, yet the precision was enhanced by 3.84 percentage points, and the parameter count was reduced by approximately 11.2%, thus facilitating the deployment. With an inference time of 39.6 FPS(frames per second), the improved model outperformed the baseline networks, including YOLOv5s, YOLOv8, and YOLOv4, thus demonstrating the high efficiency and robustness. Finally, 54 target points were measured with an average positioning error of 0.440 4 m between the center points of the targets and the anchor boxes. In summary, the finding can offer a better trade-off between speed and accuracy, in order to realize the rapid and efficient identification of gap distributions across different communities. Additionally, the precise localization requirements were fully met for the drone-based reseeding in grassland gaps. Thus, robust technical support was provided for the subsequent vegetation restoration and reconstruction in degraded grasslands. © 2024 Chinese Society of Agricultural Engineering. All rights reserved.
引用
收藏
页码:203 / 212
页数:9
相关论文
共 35 条
  • [1] ZHENG J H, WANG Q, YUAN S, Et al., Soil deterioration due to long-term grazing of desert-steppe promotes stress-tolerant ecological strategies in plants, Science of the Total Environment, 907, (2024)
  • [2] LIU Yongjie, YANG Qin, Research progress and prospect of restoration of degraded grassland on the Tibetan Plateau, Chinese Journal of Grassland, 45, 10, pp. 131-143, (2023)
  • [3] ZHANG T, BI Y, XUAN C Z., Convolutional transformer attention network with few-shot learning for grassland degradation monitoring using UAV hyperspectral imagery, International Journal of Remote Sensing, 45, 6, pp. 2109-2135, (2024)
  • [4] ZHANC X M, JOHNSTON E R, BARBERAN A, Et al., Decreased plant productivity resulting from plant group removal experiment constrains soil microbial functional diversity, Clobal Change Biology, 23, 10, pp. 4318-4332, (2017)
  • [5] SHANG C W, WU T, HUANC G L, Et al., Weak sustainability is not sustainable: socioeconomic and environmental assessment of Inner Mongolia for the past three decades, Resources, Conservation and Recycling, 141, pp. 243-252, (2019)
  • [6] ZHANG Yingjun, ZHOU Jiqiong, YANG Gaowen, Et al., Theory and practice of no-tillage reseeding and restoration of vegetation in degraded grassland, Scientific Bulletin, 65, 16, pp. 1546-1555, (2020)
  • [7] ESKANDARI R, MAHDIANPARI M, MOHAMMADIMANESH F, Et al., Meta-analysis of unmanned aerial vehicle(UAV)imagery for agro-environmental monitoring using machine learning and statistical models, Remote Sensing, 12, 21, (2020)
  • [8] LAM O H Y, DOGOTARI M, PRUM M, Et al., An open source workflow for weed mapping in native grassland using unmanned aerial vehicle: Using Rumex obtusifolius as a case study, European Journal of Remote Sensing, 54, 4, pp. 1-18, (2020)
  • [9] WANG Cuanjian, JIANG Honghong, YIN Xiaojun, Et al., Evaluation for natural grassland utilization intensity based on GPS and UAV remote sensing for grassland biomass inversion, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 34, 19, pp. 82-87, (2018)
  • [10] WANG Wenjin, YOU Ziyi, SHAO Lijiang, Et al., Recognition of dead pine trees using YOLOv5 by super-resolution reconstruction, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 39, 5, pp. 137-145, (2023)