Real-time identification, localization, and grading method for navel oranges based on RGB-D camera

被引:0
|
作者
Liu D. [1 ]
Zhu L. [1 ]
Ji W. [2 ]
Lian Y. [1 ]
机构
[1] School of Civil and Surveying Engineering, Jiangxi University of Technology, Ganzhou
[2] State Key Laboratory of Remote Sensing Science, Beijing Normal University, Beijing
关键词
depth point cloud; image recognition; instance segmentation; least-squares fitting; RANSAC; RGB-D; robot; YOLACT;
D O I
10.11975/j.issn.1002-6819.2022.14.018
中图分类号
学科分类号
摘要
The navel orange planting area in Gannan region of China has reached the first in the world, with an annual output of one million tons. However, most navel oranges were picked manually with high labor intensity and low efficiency. It was very urgent to introduce intelligent picking robots. In the process of navel orange picking, the picking robot needed to obtain the spatial position and fruit size information of navel orange fruit in real time to achieve efficient and rapid intelligent grading picking. This study proposed an algorithm framework-OrangePointSge based on the improved YOLACT model and the least-square sphere fitting method. The real-time instance segmentation algorithm YOLACT generated an instance mask and cropped the registered depth image to obtain the fruit depth point cloud. Then, the least-square method was used to fit the navel orange fruit shape, and the centroid coordinates and radius under the camera coordinate system were obtained. Firstly, the RGB-D data of navel orange fruit was collected by Microsoft's latest consumer-level depth camera (Azure Kinect DK), and the navel orange fruit instance segmentation dataset and enhancement dataset were established and free to open source. The dataset contained 2 178 images, including 8 712 samples. Among those samples, the number of non-occluded fruit was 4 682, and the number of slightly occluded fruit was 4 030. Then the YOLACT algorithm was improved to adapt to the detection of navel orange fruit, and the instance mask of the fruit was output. We modified the original ResNet (deep residual network) backbone network to HRNet (High-Resolution Net) to simplify the model, and used the hrnet_w48 structure for feature extraction to improve the model detection accuracy. We also optimized the non-maximum suppression process to improve the detection speed. At the same time, several groups of comparative tests were also set up, including the performance of YOLACT with different backbone networks and the performance of different algorithms in different scenarios. To improve the generalization ability of the training model, we carried out a series of enhancement algorithms for color images, mainly including scale change, color change, brightness change, and adding noise. For long-range small targets, we used oversampling method for enhancement training to improve their detection accuracy. Therefore, the recognition speed and mask accuracy of navel orange were higher than other algorithms. The detection speed was 44.63 frames/s and the average accuracy was 31.15%. The mask generated by improved YOLACT will be cropped with the depth image registered with the color image to obtain the depth information of each fruit to generate the fruit depth point cloud from the depth image. The least-square method was used to fit the point cloud to obtain the centroid coordinates and radius of the fruit, which were used to guide the robot to carry out grading picking. In addition, we added the comparison with RANSAC (Random Sample Consensus) algorithm. The results show that the least-square method is faster than RANSAC while ensuring the appropriate accuracy. When the number of point clouds was 1 400-2 000, the fitting time was 1.99 ms, the positioning error was 0.49 cm, the fitting radius root mean square error was 0.43 cm, and the volume root mean square error was 52.6 mL. Through the experiment of setting different distances on the fitting accuracy, it was found that when the number of point clouds was more than 800 and the distance was less than 1.0 m, the positioning error was controlled within 0.46 cm. Finally, by introducing parallel computing and combining the above two processing flows, the overall processing speed of OrangePointSge was 29.4 frames/s, which can better balance the accuracy and speed. Thus, the proposed algorithm is conducive to practical application and engineering deployment. © 2022 Chinese Society of Agricultural Engineering. All rights reserved.
引用
收藏
页码:154 / 165
页数:11
相关论文
共 32 条
  • [1] 4, pp. 149-152, (2015)
  • [2] Zheng Taixiong, Jiang Mingzhe, Feng Mingchi, Vision based target recognition and location for picking robot: A review, Chinese Journal of Scientific Instrument, 42, 9, pp. 28-51, (2021)
  • [3] Wang Dandan, Song Huaibo, He Dongjian, Research advance on vision system of apple picking robot, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 33, 10, pp. 59-69, (2017)
  • [4] Wang Hui, Mao Wenhua, Liu Gang, Et al., Identification and location system of multi-operation apple robot based on vision combination, Transactions of The Chinese Society of Agricultural Machinery, 43, 12, pp. 165-170, (2012)
  • [5] Zhang Qin, Chen Jianmin, Li Bin, Et al., Method for recognizing and locating tomato cluster picking points based on RGB-D information fusion and target detection, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 37, 18, pp. 143-152, (2021)
  • [6] Sun Baoxia, Zheng Zhenhui, Hu Wenxin, Et al., Real-time detection and localization method of longan based on RGB-D, Journal of Forestry Engineering, 7, 3, pp. 150-157, (2022)
  • [7] Fang Jianjun, Present situation and development of mobile harvesting robot, Transactions of the Chinese Society of Agricultural Engineering (Transactions of the CSAE), 20, 2, pp. 273-278, (2004)
  • [8] Li Zhipeng, Luo Yufeng, Shi Zhixin, Et al., Design and research of manipulator for naval orange harvesting, Machine Design & Research, 36, 3, pp. 188-194, (2020)
  • [9] Mai Chunyan, Zheng Lihua, Sun Hong, Et al., Research on 3D reconstruction of fruit tree and fruit recognition and location method based on RGB-D camera, Transactions of The Chinese Society for Agricultural Machinery, 46, pp. 35-40, (2015)
  • [10] Li Han, Tao Hanxiao, Cui Lihao, Et al., Recognition and localization method of tomato based on RGB-D image and improved SOM-K-means algorithm, Transac-tions of The Chinese Society for Agricultural Machinery, 52, 1, pp. 23-29, (2021)