Integrating Color Information and Multi-Scale Geometric Features for Point Cloud Semantic Segmentation

被引:0
|
作者
Zhang H. [1 ]
Xu R. [1 ]
Zheng N. [1 ]
Hao M. [1 ]
Liu D. [3 ]
Shi W. [2 ]
机构
[1] School of Environment and Spatial Informatics, China University of Mining and Technology, Xuzhou
[2] Department of Land Surveying and Geo-informatics, The Hong Kong Polytechnic University
[3] Geomatics Center of Guizhou Province, Guiyang
基金
中国国家自然科学基金;
关键词
geometric feature extraction; geometric properties; large-scale point clouds; multi-scale feature fusion; normal vector; relative positional relationship; semantic segmentation;
D O I
10.12082/dqxxkx.2024.240014
中图分类号
学科分类号
摘要
Large outdoor point clouds have rich spatial structures and are one of the important means of obtaining geographic information. They have broad application prospects in fields such as autonomous driving, robot navigation, and 3D reconstruction. Due to its inherent irregularity, complex geometric structural features, and significant changes in land scale, the accuracy of point cloud segmentation remains a huge challenge. At present, most point cloud segmentation methods only extract features based on the original 3D coordinates and color information of point cloud data and have not fully explored the information contained in point cloud data with rich spatial information, especially the problem of insufficient utilization of geometric and color information in large- scale point clouds. In order to effectively address the aforementioned issues, this paper introduces the CMGF- Net, a method for semantic segmentation of point clouds that effectively integrates color information and multi- scale geometric features. In this network, dedicated modules are designed for extracting geometric feature information and semantic feature information. In the geometric feature information extraction path, to fully leverage the geometric characteristics of point cloud data, two feature extraction modules are designed: the Relative Position Feature (RPF) extraction module and the Local Geometry Properties (LGP) extraction module, both focusing on the characteristics of the local neighborhood. In the RPF module, spatial normal information of the 3D point cloud and relative spatial distances are utilized to extract the relative positional relationships between neighboring points and the central point. The LGP module exploits the unique performance characteristics of point cloud geometric properties across different terrains, integrating geometric attribute features from the local region. Subsequently, the designed Local Geometric Feature Fusion module (LGF) combines the extracted feature information from the RPF and LGP modules, yielding fused geometric feature information. Furthermore, to learn multi- scale geometric features from the point cloud, CMGF- Net conducts geometric feature extraction at different scales within the network layers. Eventually, the extracted geometric features are hierarchically fused with semantically extracted features based on color information. By extracting multi-scale geometric features and integrating semantic features, the learning ability of the network is enhanced. The experimental results show that our proposed network model achieves a mean Intersection Over Union (mIoU) of 78.2% and an Overall Accuracy (OA) of 95.0% on the Semantic3D dataset, outperforming KPConv by 3.6% and 2.1%, respectively. On the SensatUrban dataset, it achieves a mIOU of 59.2% and an OA of 93.7%. These findings demonstrate that the proposed network model, CMGF-Net, yields promising results in the segmentation of large-scale outdoor point clouds. © 2024 Science Press. All rights reserved.
引用
收藏
页码:1562 / 1575
页数:13
相关论文
共 28 条
  • [1] Su H., Maji S., Kalogerakis E., Et al., Multi-view convolutional neural networks for 3D shape recognition[C], 2015 IEEE International Conference on Computer Vision (ICCV), pp. 945-953, (2015)
  • [2] Feng Y.F., Zhang Z.Z., Zhao X.B., Et al., GVCNN: Group-view convolutional neural networks for 3D shape recognition[C], 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 264-272, (2018)
  • [3] Charles R.Q., Hao S., Mo K.C., Et al., PointNet: Deep learning on point sets for 3D classification and segmentation [C], 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 77-85, (2017)
  • [4] Qi C.R., Yi L., Su H., Et al., PointNet++: Deep hierarchical feature learning on point sets in a metric space[C], Proceedings of the 31st International Conference on Neural Information Processing Systems, pp. 5105-5114, (2017)
  • [5] Zhou Y., Tuzel O., VoxelNet: End-to-end learning for point cloud based 3D object detection[C], 2018 IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 4490-4499, (2018)
  • [6] Tchapmi L., Choy C., Armeni I., Et al., SEGCloud: Semantic segmentation of 3D point clouds[C], 2017 International Conference on 3D Vision (3DV), pp. 537-547, (2017)
  • [7] Riegler G., Ulusoy A.O., Geiger A., OctNet: Learning deep 3D representations at high resolutions[C], 2017 IEEE Conference on Computer Vision and Pattern Recognition (CVPR), pp. 6620-6629, (2017)
  • [8] Li Y., Bu R., Sun M., Et al., Pointcnn: convolution on x-transformed points[C], Advances in neural information processing systems, (2018)
  • [9] Thomas H., Qi C.R., Deschaud J.E., Et al., KPConv: Flexible and deformable convolution for point clouds[C], 2019 IEEE/CVF International Conference on Computer Vision (ICCV), pp. 6410-6419, (2019)
  • [10] Boulch A., ConvPoint: Continuous convolutions for point cloud processing[EB/OL], (2019)