Lightweight monocular absolute depth estimation based on attention mechanism

被引:1
|
作者
Jin, Jiayu [1 ,2 ]
Tao, Bo [1 ]
Qian, Xinbo [2 ,3 ]
Hu, Jiaxin [3 ]
Li, Gongfa [4 ]
机构
[1] Wuhan Univ Sci & Technol, Key Lab Met Equipment & Control Technol, Minist Educ, Wuhan, Peoples R China
[2] Wuhan Univ Sci & Technol, Hubei Key Lab Mech Transmiss & Mfg Engn, Wuhan, Peoples R China
[3] Wuhan Univ Sci & Technol, Precis Mfg Inst, Wuhan, Peoples R China
[4] Wuhan Univ Sci & Technol, Res Ctr Biomimet Robot & Intelligent Measurement &, Wuhan, Peoples R China
关键词
lightweight network; deep learning; monocular depth estimation; channel attention; self-supervised;
D O I
10.1117/1.JEI.33.2.023010
中图分类号
TM [电工技术]; TN [电子技术、通信技术];
学科分类号
0808 ; 0809 ;
摘要
To solve the problem of obtaining a higher accuracy at the expense of redundant models, we propose a network architecture. We utilize a lightweight network that retains the high-precision advantage of the transformer and effectively combines it with convolutional neural network. By greatly reducing the training parameters, this approach achieves high precision, making it well suited for deployment on edge devices. A detail highlight module (DHM) is added to effectively fuse information from multiple scales, making the depth of prediction more accurate and clearer. A dense geometric constraints module is introduced to recover accurate scale factors in autonomous driving without additional sensors. Experimental results demonstrate that our model improves the accuracy from 98.1% to 98.3% compared with Monodepth2, and the model parameters are reduced by about 80%.
引用
收藏
页数:13
相关论文
共 50 条
  • [31] Unsupervised Monocular Depth Estimation With Channel and Spatial Attention
    Wang, Zhuping
    Dai, Xinke
    Guo, Zhanyu
    Huang, Chao
    Zhang, Hao
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (06) : 7860 - 7870
  • [32] Unsupervised Monocular Depth Estimation and Visual Odometry Based on Generative Adversarial Network and Self-attention Mechanism
    Ye X.
    He Y.
    Ru S.
    Jiqiren/Robot, 2021, 43 (02): : 203 - 213
  • [33] CATNet: Convolutional attention and transformer for monocular depth estimation
    Tang, Shuai
    Lu, Tongwei
    Liu, Xuanxuan
    Zhou, Huabing
    Zhang, Yanduo
    PATTERN RECOGNITION, 2024, 145
  • [34] Illumination Insensitive Monocular Depth Estimation Based on Scene Object Attention and Depth Map Fusion
    Wen, Jing
    Ma, Haojiang
    Yang, Jie
    Zhang, Songsong
    PATTERN RECOGNITION AND COMPUTER VISION, PRCV 2023, PT X, 2024, 14434 : 358 - 370
  • [35] CNNapsule: A Lightweight Network with Fusion Features for Monocular Depth Estimation
    Wang, Yinchu
    Zhu, Haijiang
    Liu, Mengze
    ARTIFICIAL NEURAL NETWORKS AND MACHINE LEARNING - ICANN 2021, PT I, 2021, 12891 : 507 - 518
  • [36] LW-Net: A Lightweight Network for Monocular Depth Estimation
    Feng, Cheng
    Zhang, Congxuan
    Chen, Zhen
    Li, Ming
    Chen, Hao
    Fan, Bingbing
    IEEE ACCESS, 2020, 8 : 196287 - 196298
  • [37] Self-Supervised Monocular Depth Estimation for Traffic Scenes Based on Dual Attention Mechanism and Adaptive Cost Volume
    Wu G.
    Liu W.
    Hu J.
    Cheng S.
    Yang W.-X.
    Sun L.-K.
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2024, 52 (05): : 1670 - 1678
  • [38] RENA-Depth: toward recursion representation enhancement in neighborhood attention guided lightweight self-supervised monocular depth estimation
    Yang, Chaochao
    Lu, Yuanyao
    Qiu, Yongsheng
    Wang, Yuantao
    OPTICAL ENGINEERING, 2024, 63 (08)
  • [39] Monocular depth estimation with multi-view attention autoencoder
    Geunho Jung
    Sang Min Yoon
    Multimedia Tools and Applications, 2022, 81 : 33759 - 33770
  • [40] Patch-Wise Attention Network for Monocular Depth Estimation
    Lee, Sihaeng
    Lee, Janghyeon
    Kim, Byungju
    Yi, Eojindl
    Kim, Junmo
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 1873 - 1881