Using full-scale feature fusion for self-supervised indoor depth estimation

被引:0
|
作者
Cheng, Deqiang [1 ]
Chen, Junhui [1 ]
Lv, Chen [1 ]
Han, Chenggong [1 ]
Jiang, He [1 ]
机构
[1] China Univ Min & Technol, Sch Informat & Control Engn, Xuzhou 221116, Peoples R China
基金
中国国家自然科学基金;
关键词
Monocular depth estimation; Feature fusion; Self-supervised; Indoor scenes; ResNeSt;
D O I
10.1007/s11042-023-16581-6
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Monocular depth estimation is a crucial task in computer vision, and self-supervised algorithms are gaining popularity due to their independence from expensive ground truth supervision. However, current self-supervised algorithms may not provide accurate estimation and may suffer from distorted boundaries when applied to indoor scenes. Combining multi-scale features is an important research direction in image segmentation to achieve accurate estimation and resolve boundary distortion. However, there are few studies on indoor self-supervised algorithms in this regard. To solve this issue, we propose a novel full-scale feature information fusion approach that includes a full-scale skip-connection and a full-scale feature fusion block. This approach can aggregate the high-level and low-level information of all scale feature maps during the network's encoding and decoding process to compensate for the network's loss of cross-layer feature information. The proposed full-scale feature fusion improves accuracy and reduces the decoder parameters. To fully exploit the superiority of the full-scale feature fusion module, we replace the encoder backbone from ResNet with the more advanced ResNeSt. Combining these two methods results in a significant improvement in prediction accuracy. We have extensively evaluated our approach on the indoor benchmark datasets NYU Depth V2 and ScanNet. Our experimental results demonstrate that our method outperforms existing algorithms, particularly on NYU Depth V2, where our precision is raised to 83.8%.
引用
收藏
页码:28215 / 28233
页数:19
相关论文
共 50 条
  • [31] MonoIndoor: Towards Good Practice of Self-Supervised Monocular Depth Estimation for Indoor Environments
    Ji, Pan
    Li, Runze
    Bhanu, Bir
    Xu, Yi
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 12767 - 12776
  • [32] GasMono: Geometry-Aided Self-Supervised Monocular Depth Estimation for Indoor Scenes
    Zhao, Chaoqiang
    Poggi, Matteo
    Tosi, Fabio
    Zhou, Lei
    Sun, Qiyu
    Tang, Yang
    Mattoccia, Stefano
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 16163 - 16174
  • [33] Self-Supervised Monocular Depth Estimation Using Hybrid Transformer Encoder
    Hwang, Seung-Jun
    Park, Sung-Jun
    Baek, Joong-Hwan
    Kim, Byungkyu
    IEEE SENSORS JOURNAL, 2022, 22 (19) : 18762 - 18770
  • [34] MonoDiffusion: Self-Supervised Monocular Depth Estimation Using Diffusion Model
    Shao, Shuwei
    Pei, Zhongcai
    Chen, Weihai
    Sun, Dingchi
    Chen, Peter C. Y.
    Li, Zhengguo
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (04) : 3664 - 3678
  • [35] HI-Net: Boosting Self-Supervised Indoor Depth Estimation via Pose Optimization
    Wu, Guanghui
    Li, Kunhong
    Wang, Longguang
    Hu, Ruizhen
    Guo, Yulan
    Chen, Zengping
    IEEE ROBOTICS AND AUTOMATION LETTERS, 2023, 8 (01) : 224 - 231
  • [36] Joint Self-Supervised Monocular Depth Estimation and SLAM
    Xing, Xiaoxia
    Cai, Yinghao
    Lu, Tao
    Yang, Yiping
    Wen, Dayong
    2022 26TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION (ICPR), 2022, : 4030 - 4036
  • [37] Learn to Adapt for Self-Supervised Monocular Depth Estimation
    Sun, Qiyu
    Yen, Gary G.
    Tang, Yang
    Zhao, Chaoqiang
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (11) : 15647 - 15659
  • [38] Towards Keypoint Guided Self-supervised Depth Estimation
    Bartol, Kristijan
    Bojanic, David
    Petkovic, Tomislav
    Pribanic, Tomislav
    Donoso, Yago
    VISAPP: PROCEEDINGS OF THE 15TH INTERNATIONAL JOINT CONFERENCE ON COMPUTER VISION, IMAGING AND COMPUTER GRAPHICS THEORY AND APPLICATIONS, VOL 4: VISAPP, 2020, : 583 - 589
  • [39] Semantically guided self-supervised monocular depth estimation
    Lu, Xiao
    Sun, Haoran
    Wang, Xiuling
    Zhang, Zhiguo
    Wang, Haixia
    IET IMAGE PROCESSING, 2022, 16 (05) : 1293 - 1304
  • [40] Self-supervised recurrent depth estimation with attention mechanisms
    Makarov, Ilya
    Bakhanova, Maria
    Nikolenko, Sergey
    Gerasimova, Olga
    PEERJ COMPUTER SCIENCE, 2022, 8