High-precision target ranging in complex orchard scenes by utilizing semantic segmentation results and binocular vision

被引:8
|
作者
Wen, Yu [1 ]
Xue, Jinlin [1 ]
Sun, Han [1 ]
Song, Yue [1 ]
Lv, Pengfei [1 ]
Liu, Shaohua [1 ]
Chu, Yangyang [1 ]
Zhang, Tianyu [1 ]
机构
[1] Nanjing Agr Univ, Coll Engn, Nanjing 210031, Peoples R China
关键词
Orchard; Deep learning; Semantic segmentation; Binocular vision; Attention mechanism; Feature fusion; AGRICULTURE;
D O I
10.1016/j.compag.2023.108440
中图分类号
S [农业科学];
学科分类号
09 ;
摘要
The automation of orchard production is increasingly relying on robotics, driven by the advancements in artificial intelligence technology. However, accurately comprehending semantic information and precisely locating various targets within orchard environments remain challenges. Current research often relies on expensive multisensor fusion techniques or vision-only approaches that yield inadequate segmentation outcomes for perceiving orchard surroundings. To address these issues, this article proposes a novel approach for target ranging in complex orchard scenes, leveraging semantic segmentation results. The article introduces the MsFF-Segformer model, which employs multi-scale feature fusion to generate high-precision semantic segmentation images. The model incorporates the MiT-B0 encoder, which utilizes a pure attention mechanism, and the MsFF decoder, specifically designed for multi-scale feature fusion. The MsFF decoder includes the AFAM module to effectively align features of adjacent scales. Additionally, the channel attention module and depth separable convolution module are introduced to reduce model parameter size and obtain feature vectors with rich semantic levels, enhancing the segmentation performance of multi-scale targets in orchards. Based on the accurate semantic segmentation outcomes in orchard environments, this study introduces a novel approach named TPDMR that integrates binocular vision to estimate the distances of various objects within orchards. Firstly, the process involves matching the semantic category matrix with the depth information matrix. Subsequently, the depth information array that represents the target category is obtained, and any invalid depth information is filtered out. Finally, the average depth of the target is calculated. Evaluation of the MsFF-Segformer model on a self-made orchard dataset demonstrates superior performance compared to U-net and other models, achieving a Mean Intersection over Union (MIoU) of 86.52 % and a Mean Pixel Accuracy (MPA) of 94.05 %. The parameters and prediction time for a single frame are 15.1 M and 0.019 s, respectively. These values are significantly lower than those of U-net, Deeplabv3+, and Hrnet models, with reductions of 84.1 %, 32.5 %, 5.9 % and 69.4 %, 59.7 %, 64.2 % respectively. The TPDMR method demonstrates a high level of accuracy and stability in target ranging, with a ranging error of less than 6 % across all targets. Furthermore, the overall algorithm runtime is estimated to be approximately 0.8 s, indicating efficient performance.
引用
收藏
页数:10
相关论文
共 35 条
  • [21] Multi-Level Branch Cross-Scale Fusion Network for High-Precision Semantic Segmentation in Complex Remote Sensing Environments
    Zeng, Junying
    Deng, Senyao
    Qin, Chuanbo
    Zhai, Yikui
    Jia, Xudong
    Gu, Yajin
    Xu, Jiahua
    LASER & OPTOELECTRONICS PROGRESS, 2025, 62 (04)
  • [22] Vision-based target localization and online error correction for high-precision robotic drilling
    Maghami, Ali
    Khoshdarregi, Matt
    ROBOTICA, 2024, 42 (09) : 3019 - 3043
  • [23] High-precision and robust localization system for mobile robots in complex and large-scale indoor scenes
    Wang, Jibo
    Li, Chengpeng
    Li, Bangyu
    Pang, Chenglin
    Fang, Zheng
    INTERNATIONAL JOURNAL OF ADVANCED ROBOTIC SYSTEMS, 2021, 18 (05)
  • [24] High-Precision Positioning and Rotation Angle Estimation for a Target Pallet Based on BeiDou Navigation Satellite System and Vision
    Meng, Deqiang
    Ren, Yufei
    Yu, Xinli
    Yin, Xiaoxv
    Wang, Wenming
    Men, Junhui
    SENSORS, 2024, 24 (16)
  • [25] High-precision six-degree-of-freedom pose measurement and grasping system for large-size object based on binocular vision
    Wan, Guoyang
    Li, Fudong
    Zhu, Wenjun
    Wang, Guofeng
    SENSOR REVIEW, 2020, 40 (01) : 71 - 80
  • [26] High-precision robotic microcontact printing (R-μCP) utilizing a vision guided selectively compliant articulated robotic arm
    McNulty, Jason D.
    Klann, Tyler
    Sha, Jin
    Salick, Max
    Knight, Gavin T.
    Turng, Lih-Sheng
    Ashton, Randolph S.
    LAB ON A CHIP, 2014, 14 (11) : 1923 - 1930
  • [27] Optimized multi-line laser calibration based on innovative 3D target for high-precision vision systems
    Yang, Ziyi
    Yang, Pei
    Wu, Qiang
    Zhang, Jin
    Xia, Haojie
    MEASUREMENT SCIENCE AND TECHNOLOGY, 2025, 36 (02)
  • [28] Efficient multi-modal high-precision semantic segmentation from MLS point cloud without 3D annotation
    Wang, Yuan
    Sun, Pei
    Chu, Wenbo
    Li, Yuhao
    Chen, Yiping
    Lin, Hui
    Dong, Zhen
    Yang, Bisheng
    He, Chao
    INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2024, 135
  • [29] Research on Real-time and High-precision Positioning Method of Ground Target through UAV Stereo Vision and Spatial Information Fusion
    Wang, Ping
    Luo, Xianquan
    Junwei, Lv
    RECENT ADVANCES IN ELECTRICAL & ELECTRONIC ENGINEERING, 2023, 16 (03) : 211 - 223
  • [30] Conceptual Validation of High-Precision Fish Feeding Behavior Recognition Using Semantic Segmentation and Real-Time Temporal Variance Analysis for Aquaculture
    Kong, Han
    Wu, Junfeng
    Liang, Xuelan
    Xie, Yongzhi
    Qu, Boyu
    Yu, Hong
    BIOMIMETICS, 2024, 9 (12)