Depth completion method based on multi-guided structure-aware networks

被引:0
|
作者
Sun H. [1 ,2 ]
Jin Y.-Q. [1 ,2 ]
Zhang W.-A. [1 ,2 ]
Fu M.-L. [1 ,2 ]
机构
[1] College of Information Engineering, Zhejiang University of Technology, Hangzhou
[2] Zhejiang Provincial United Key Laboratory of Embedded Systems, Hangzhou
来源
Kongzhi yu Juece/Control and Decision | 2024年 / 39卷 / 02期
关键词
adaptive receptive field; depth completion; image guide filtering; multi-sensor fusion; sparse scene; structure awareness;
D O I
10.13195/j.kzyjc.2022.0759
中图分类号
学科分类号
摘要
Aiming at the problem of sparse depth information observation in 3D scenes, this paper proposes a multi-guided structure-aware network model fused with color images to complement the sparse depth. Using the mapping relationship between the 3D plane normal vector and the scene gradient information, we design a two-branch backbone network framework and combine image features and geometric features for depth prediction to fully extract the feature representation of spatial location information. Secondly, considering the structural differences of different objects in large-scale scenes, a network channel attention mechanism is designed. A structure-aware module with an adaptive receptive field is used to characterize information at different scales. Finally, in the process of network upsampling, the predicted sub-depth maps are filtered and the edge details of objects are repaired with the guidance of images of different sizes. The experimental results on public datasets show that the designed depth completion algorithm can obtain accurate dense depth. At the same time, through the in-depth evaluation of two downstream sensing tasks, the results show that the propsed method can effectively improve the effect of other sensing tasks. © 2024 Northeast University. All rights reserved.
引用
收藏
页码:401 / 410
页数:9
相关论文
共 35 条
  • [1] Gu J Q, Xiang Z Y, Ye Y W, Et al., DenseLiDAR: A real-time pseudo dense depth guided depth completion network, IEEE Robotics and Automation Letters, 6, 2, pp. 1808-1815, (2021)
  • [2] Vu T H, Jain H, Bucher M, Et al., DADA: Depth-aware domain adaptation in semantic segmentation, IEEE/CVF International Conference on Computer Vision, pp. 7363-7372, (2020)
  • [3] Borghi G, Fabbri M, Vezzani R, Et al., Face-from-depth for head pose estimation on depth images, IEEE Transactions on Pattern Analysis and Machine Intelligence, 42, 3, pp. 596-609, (2020)
  • [4] Zheng T X, Huang S, Li Y F, Et al., Key techniques for vision based 3D reconstruction: A review, Acta Automatica Sinica, 46, 4, pp. 631-652, (2020)
  • [5] Uhrig J, Schneider N, Schneider L, Et al., Sparsity invariant CNNs, International Conference on 3D Vision, pp. 11-20, (2018)
  • [6] Zuo Y F, Wu Q, Zhang J, Et al., Explicit edge inconsistency evaluation model for color-guided depth map enhancement, IEEE Transactions on Circuits and Systems for Video Technology, 28, 2, pp. 439-453, (2018)
  • [7] Xue H Y, Zhang S M, Cai D., Depth image inpainting: Improving low rank matrix completion with low gradient regularization, IEEE Transactions on Image Processing: A Publication of the IEEE Signal Processing Society, 26, 9, pp. 4311-4320, (2017)
  • [8] Ma F C, Carlone L, Ayaz U, Et al., Sparse depth sensing for resource-constrained robots, The International Journal of Robotics Research, 38, 8, pp. 935-980, (2019)
  • [9] Zhao Y M, Bai L, Zhang Z M, Et al., A surface geometry model for LiDAR depth completion, IEEE Robotics and Automation Letters, 6, 3, pp. 4457-4464, (2021)
  • [10] Ren M Y, Pokrovsky A, Yang B, Et al., SBNet: Sparse blocks network for fast inference, IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 8711-8720, (2018)