3D Vision-Based Picking System with Instance Segmentation Network and Iterative Optimization Method

被引:0
|
作者
Wang D. [1 ]
Yan Y. [1 ]
Zhou G. [1 ]
Li Y. [1 ]
Liu C. [1 ]
Lin L. [1 ]
Chen Q. [1 ]
机构
[1] College of Electronics and Information Engineering, Tongji University, Shanghai
来源
Jiqiren/Robot | 2019年 / 41卷 / 05期
关键词
3D object detection; Instance segmentation; Picking system; Pose estimation; RGB-D; Texture-less object;
D O I
10.13973/j.cnki.robot.180806
中图分类号
学科分类号
摘要
A workpiece recognition and picking system based on instance segmentation network and iterative optimization method is proposed for object detection and pose estimation of scattered and stacked texture-less industrial objects. This system consists of three modules, including image acquisition module, target detection module and pose estimation module. In image acquisition module, a dual RGB-D (RGB-depth) camera structure is designed to get higher quality depth data by merging three depth images. The target detection module modifies the instance segmentation network Mask R-CNN (regionbased convolutional neural network). The modified network takes RGB images and HHA (horizontal disparity, height above ground, angle with gravity) features containing three-dimensional information as input, and adds STN (spatial transformer network) modules inside to improve the segmentation performance of texture-less objects. Then the module can combine point cloud information to obtain the target point cloud. On this basis, the improved 4PCS (4-points congruent set) algorithm and ICP (iterative closest point) algorithm are used in pose estimation module to match the segmented point cloud with the target model and fine-tune the pose, and thus the final result of pose estimation is obtained. The robots accomplish picking action according to the estimated pose. The experiment results on our workpiece dataset and the actual picking system indicate that the proposed method can achieve fast target recognition and pose estimation for scattered and stacked objects with different shapes and less textures. Its performance can meet the requirements of practical applications with 1 mm position error and 1° angle error. © 2019, Science Press. All right reserved.
引用
收藏
页码:637 / 648
页数:11
相关论文
共 30 条
  • [1] Hinterstoisser S., Lepetit V., Ilic S., Et al., Model based training, detection and pose estimation of texture-less 3D objects in heavily cluttered scenes, Asian Conference on Computer Vision, pp. 548-562, (2012)
  • [2] Duff D.J., Morwald T., Stolkin R., Et al., Physical simulation for monocular 3D model based tracking, IEEE International Conference on Robotics and Automation, pp. 5218-5225, (2011)
  • [3] Guo Y., Sohel F., Bennamoun M., Et al., Rotational projection statistics for 3D local surface description and object recognition, International Journal of Computer Vision, 105, 1, pp. 63-86, (2013)
  • [4] Drost B., Ulrich M., Navab N., Et al., Model globally, match locally: Efficient and robust 3D object recognition, IEEE Confer Ence on Computer Vision and Pattern Recognition, pp. 998-1005, (2010)
  • [5] Wu X.R., Huang G.M., Sun L.N., Fast visual identification and location algorithm for industrial sorting robots based on deep learning, Robot, 38, 6, pp. 711-719, (2016)
  • [6] Du X.D., Cai Y.H., Lu T., Et al., A robotic grasping method based on deep learning, Robot, 39, 6, (2017)
  • [7] Xia J., Qian K., Ma X.D., Et al., Fast planar grasp pose detection for robot based on cascaded deep convolutional neural networks, Robot, 40, 6, pp. 794-802, (2018)
  • [8] Zeng A., Song S., Yu K.T., Et al., Robotic pick-and-place of novel objects in clutter with multi-affordance grasping and cross-domain image matching, IEEE International Conference on Robotics and Automation, pp. 3750-3757, (2018)
  • [9] Gupta S., Girshick R., Arbelaez P., Et al., Learning rich features from RGB-D images for object detection and segmentation, 13th European Conference on Computer Vision, pp. 345-360, (2014)
  • [10] Alexandre L.A., 3D object recognition using convolutional neural networks with transfer learning between input channels, 13th International Conference on Intelligent Autonomous Systems, pp. 888-897, (2016)