3D Instance Segmentation Using Deep Learning on RGB-D Indoor Data

被引:3
|
作者
Yasir, Siddiqui Muhammad [1 ]
Sadiq, Amin Muhammad [2 ]
Ahn, Hyunsik [3 ]
机构
[1] Tongmyong Univ, Dept Robot Syst Engn, Busan 48520, South Korea
[2] Univ Cent Punjab, Dept Informat & Technol, Lahore, Pakistan
[3] Tongmyong Univ, Dept Elect Engn, Busan 48520, South Korea
来源
CMC-COMPUTERS MATERIALS & CONTINUA | 2022年 / 72卷 / 03期
关键词
Instance segmentation; 3D object segmentation; deep learning; point cloud coordinates;
D O I
10.32604/cmc.2022.025909
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
3D object recognition is a challenging task for intelligent and robot systems in industrial and home indoor environments. It is critical for such systems to recognize and segment the 3D object instances that they encounter on a frequent basis. The computer vision, graphics, and machine learning fields have all given it a lot of attention. Traditionally, 3D segmentation was done with hand-crafted features and designed approaches that didn't achieve acceptable performance and couldn't be generalized to large-scale data. Deep learning approaches have lately become the preferred method for 3D segmentation challenges by their great success in 2D computer vision. However, the task of instance segmentation is currently less explored. In this paper, we propose a novel approach for efficient 3D instance segmentation using red green blue and depth (RGB-D) data based on deep learning. The 2D region based convolutional neural networks (Mask R-CNN) deep learning model with point based rending module is adapted to integrate with depth information to recognize and segment 3D instances of objects. In order to generate 3D point cloud coordinates (x, y, z), segmented 2D pixels (u, v) of recognized object regions in the RGB image are merged into (u, v) points of the depth image. Moreover, we conducted an experiment and analysis to compare our proposed method from various points of view and distances. The experimentation shows the proposed 3D object recognition and instance segmentation are sufficiently beneficial to support object handling in robotic and intelligent systems.
引用
收藏
页码:5777 / 5791
页数:15
相关论文
共 50 条
  • [1] 3D mapping of indoor environments using RGB-D data
    dos Santos, Daniel Rodrigues
    Khoshelham, Kourosh
    BOLETIM DE CIENCIAS GEODESICAS, 2015, 21 (03): : 442 - 464
  • [2] RGB-D Semantic Segmentation for Indoor Modeling Using Deep Learning: A Review
    Rached, Ishraq
    Hajji, Rafika
    Landes, Tania
    RECENT ADVANCES IN 3D GEOINFORMATION SCIENCE, 3D GEOINFO 2023, 2024, : 587 - 604
  • [3] 3D-SIS: 3D Semantic Instance Segmentation of RGB-D Scans
    Hou, Ji
    Dai, Angela
    Niessner, Matthias
    2019 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2019), 2019, : 4416 - 4425
  • [4] Applications of RGB-D data for 3D reconstruction in the indoor environment
    Zou, Runyang
    Ge, Xueshi
    Wang, Geng
    2016 IEEE CHINESE GUIDANCE, NAVIGATION AND CONTROL CONFERENCE (CGNCC), 2016, : 375 - 378
  • [5] Neural Colour Correction for Indoor 3D Reconstruction Using RGB-D Data
    Madeira, Tiago
    Oliveira, Miguel
    Dias, Paulo
    SENSORS, 2024, 24 (13)
  • [6] 3D Spatial Layout Extraction of Indoor Images Using RGB-D Data
    Yapicilar, Caglar
    Arica, Nafiz
    2013 21ST SIGNAL PROCESSING AND COMMUNICATIONS APPLICATIONS CONFERENCE (SIU), 2013,
  • [7] Instance segmentation of point cloud captured by RGB-D sensor based on deep learning
    Wang, Zhengtuo
    Xu, Yuetong
    Yu, Jiongyan
    Xu, Guanhua
    Fu, Jianzhong
    Gu, Tianyi
    INTERNATIONAL JOURNAL OF COMPUTER INTEGRATED MANUFACTURING, 2021, 34 (09) : 950 - 963
  • [8] Semantic Segmentation Networks of 3D Point Clouds for RGB-D Indoor Scenes
    Wang, Ya
    Zell, Andreas
    TWELFTH INTERNATIONAL CONFERENCE ON MACHINE VISION (ICMV 2019), 2020, 11433
  • [9] 3D indoor scene modeling from RGB-D data: A survey
    Chen K.
    Lai Y.-K.
    Hu S.-M.
    Computational Visual Media, 2015, 1 (4) : 267 - 278
  • [10] 3D indoor scene modeling from RGB-D data:a survey
    Kang Chen
    Yu-Kun Lai
    Shi-Min Hu
    Computational Visual Media, 2015, 1 (04) : 267 - 278