SABV-Depth: A biologically inspired deep learning network for monocular depth estimation

被引:10
|
作者
Wang, Junfan [1 ,2 ]
Chen, Yi [1 ,2 ]
Dong, Zhekang [1 ,2 ,3 ]
Gao, Mingyu [1 ,2 ]
Lin, Huipin [1 ,2 ]
Miao, Qiheng [4 ]
机构
[1] Hangzhou Dianzi Univ, Sch Elect Informat, Hangzhou 310018, Zhejiang, Peoples R China
[2] Zhejiang Prov Key Lab Equipment Elect, Hangzhou 310018, Zhejiang, Peoples R China
[3] Zhejiang Univ, Dept Elect Engn, Hangzhou 310027, Zhejiang, Peoples R China
[4] Zhejiang Huaruijie Technol Co Ltd, Hangzhou 310051, Zhejiang, Peoples R China
关键词
Depth estimation; Biological vision; Mapping relationship; Self -attention mechanism; VISION; MODEL; CONSCIOUSNESS;
D O I
10.1016/j.knosys.2023.110301
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Monocular depth estimation makes it possible for machines to perceive the real world. The prediction performance of the depth estimation network based on deep learning will be affected due to the depth of the deep network and the locality of convolution operations. The imitation of the biological visual system and its functional structure is becoming a research hotspot. In this paper, we study the interpretability relationship between the biological visual system and the monocular depth estimation network. By concretizing the attention mechanism in biological vision, we propose a monocular depth estimation network based on the self-attention mechanism, named SABV-Depth, which can improve prediction accuracy. Inspired by the biological visual interaction mechanism, we focus on the information transfer between each module of the network and improve the information retention ability, and enable the network to output a depth map with rich object information and detailed information. Further, a decoder module with an inner-connection is proposed to recover depth maps with sharp edge contours. Our method is experimentally validated on the KITTI dataset and NYU Depth V2 dataset. The results show that compared with other works, the proposed method improves prediction accuracy. Meanwhile, the depth map has more object information and detail information, and a better edge information processing effect. (c) 2023 The Author(s). Published by Elsevier B.V. This is an open access article under the CC BY-NC-ND
引用
收藏
页数:14
相关论文
共 50 条
  • [31] Unsupervised Learning of Monocular Depth Estimation and Visual Odometry with Deep Feature Reconstruction
    Zhan, Huangying
    Garg, Ravi
    Weerasekera, Chamara Saroj
    Li, Kejie
    Agarwal, Harsh
    Reid, Ian
    2018 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR), 2018, : 340 - 349
  • [32] DCL-depth: monocular depth estimation network based on iam and depth consistency loss
    Han C.
    Lv C.
    Kou Q.
    Jiang H.
    Cheng D.
    Multimedia Tools and Applications, 2025, 84 (8) : 4773 - 4787
  • [33] Towards Interpretable Deep Networks for Monocular Depth Estimation
    You, Zunzhi
    Tsai, Yi-Hsuan
    Chiu, Wei-Chen
    Li, Guanbin
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 12859 - 12868
  • [34] Error Diagnosis of Deep Monocular Depth Estimation Models
    Chawla, Jagpreet
    Thakurdesai, Nikhil
    Godase, Anuj
    Reza, Md
    Crandall, David
    Jung, Soon-Heung
    2021 IEEE/RSJ INTERNATIONAL CONFERENCE ON INTELLIGENT ROBOTS AND SYSTEMS (IROS), 2021, : 5344 - 5349
  • [35] Sparse depth densification for monocular depth estimation
    Zhen Liang
    Tiyu Fang
    Yanzhu Hu
    Yingjian Wang
    Multimedia Tools and Applications, 2024, 83 : 14821 - 14838
  • [36] Sparse depth densification for monocular depth estimation
    Liang, Zhen
    Fang, Tiyu
    Hu, Yanzhu
    Wang, Yingjian
    MULTIMEDIA TOOLS AND APPLICATIONS, 2024, 83 (05) : 14821 - 14838
  • [37] Depth Map Decomposition for Monocular Depth Estimation
    Jun, Jinyoung
    Lee, Jae-Han
    Lee, Chul
    Kim, Chang-Su
    COMPUTER VISION - ECCV 2022, PT II, 2022, 13662 : 18 - 34
  • [38] Monocular Depth Estimation Using Information Exchange Network
    Su, Wen
    Zhang, Haifeng
    Zhou, Quan
    Yang, Wenzhen
    Wang, Zengfu
    IEEE TRANSACTIONS ON INTELLIGENT TRANSPORTATION SYSTEMS, 2021, 22 (06) : 3491 - 3503
  • [39] A NOVEL LIGHTWEIGHT NETWORK FOR FAST MONOCULAR DEPTH ESTIMATION
    Heydrich, Tim
    Yang, Yimin
    Ma, Xiangyu
    Liu, Yu
    Du, Shan
    2022 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING (ICASSP), 2022, : 2260 - 2264
  • [40] GBNet: Gradient Boosting Network for Monocular Depth Estimation
    Han, Daechan
    Choi, Yukyung
    2021 21ST INTERNATIONAL CONFERENCE ON CONTROL, AUTOMATION AND SYSTEMS (ICCAS 2021), 2021, : 342 - 346