Integrating convolutional guidance and Transformer fusion with Markov Random Fields smoothing for monocular depth estimation

被引:0
|
作者
Peng, Xiaorui [1 ]
Meng, Yu [1 ]
Shi, Boqiang [1 ]
Zheng, Chao [1 ]
Wang, Meijun [1 ]
机构
[1] Univ Sci & Technol Beijing, XueYuan Rd 30, Beijing 100083, Peoples R China
关键词
Monocular depth estimation; Intelligent transportation; Environment perception;
D O I
10.1016/j.engappai.2025.110011
中图分类号
TP [自动化技术、计算机技术];
学科分类号
0812 ;
摘要
Monocular depth estimation is a challenging and prominent problem in current computer vision research and is widely used in intelligent transportation like environment perception, navigation and localization. Accurately delineating object boundaries and ensuring smooth transitions in estimated depth images from a single image remain significant challenges. These issues place higher demands on the network's global and local feature extraction capabilities. In response, we proposed a depth estimation framework, designed to address detection accuracy and the global smooth transition of predicted depth maps. Our method introduces a novel feature decoding structure named Convolutional Guided Fusion (CoGF), which utilizes local features extracted by a convolutional neural network as a guide and fuses them with long-range dependent features extracted by a Transformer. This approach enables the model to retain both local details and global contextual information during the decoding process. To ensure global smoothness in the depth estimation results, we incorporate a smoothing strategy based on Markov Random Fields (MRF), enhancing pixel-to-pixel continuity and ensuring robust spatial consistency in the generated depth maps. Our proposed method is evaluated on current mainstream benchmarks. Experimental results demonstrate that our depth estimation method outperforms previous approaches. The code is available at https://github.com/pxrw/CGTF-Depth.git.
引用
收藏
页数:10
相关论文
共 50 条
  • [31] Feature Enhanced Fully Convolutional Networks for Monocular Depth Estimation
    Shi, Chunxiu
    Chen, Jie
    Chen, Juan
    Zhang, Zheng
    2019 IEEE INTERNATIONAL CONFERENCE ON DATA SCIENCE AND ADVANCED ANALYTICS (DSAA 2019), 2019, : 270 - 276
  • [32] Registration Fusion Using Markov Random Fields
    Gass, Tobias
    Szekely, Gabor
    Goksel, Orcun
    BIOMEDICAL IMAGE REGISTRATION (WBIR 2014), 2014, 8545 : 213 - 222
  • [33] Locally Adaptive Smoothing with Markov Random Fields and Shrinkage Priors
    Faulkner, James R.
    Minin, Vladimir N.
    BAYESIAN ANALYSIS, 2018, 13 (01): : 225 - 252
  • [34] Self-supervised multi-frame depth estimation with visual-inertial pose transformer and monocular guidance
    Wang, Xiang
    Luo, Haonan
    Wang, Zihang
    Zheng, Jin
    Bai, Xiao
    INFORMATION FUSION, 2024, 108
  • [35] Dense monocular depth estimation for stereoscopic vision based on pyramid transformer and multi-scale feature fusion
    Xia, Zhongyi
    Wu, Tianzhao
    Wang, Zhuoyan
    Zhou, Man
    Wu, Boqi
    Chan, C. Y.
    Kong, Ling Bing
    SCIENTIFIC REPORTS, 2024, 14 (01)
  • [36] Dense monocular depth estimation for stereoscopic vision based on pyramid transformer and multi-scale feature fusion
    Zhongyi Xia
    Tianzhao Wu
    Zhuoyan Wang
    Man Zhou
    Boqi Wu
    C. Y. Chan
    Ling Bing Kong
    Scientific Reports, 14
  • [37] DEPTH RECOVERY USING MARKOV RANDOM-FIELDS
    KAPOOR, S
    MUNDKUR, PY
    DESAI, UB
    SADHANA-ACADEMY PROCEEDINGS IN ENGINEERING SCIENCES, 1993, 18 : 17 - 29
  • [38] Plane2Depth: Hierarchical Adaptive Plane Guidance for Monocular Depth Estimation
    Liu, Li
    Zhu, Ruijie
    Deng, Jiacheng
    Song, Ziyang
    Yang, Wenfei
    Zhang, Tianzhu
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2025, 35 (02) : 1136 - 1149
  • [39] Monocular Depth Estimation Based on Multi-Scale Depth Map Fusion
    Yang, Xin
    Chang, Qingling
    Liu, Xinglin
    He, Siyuan
    Cui, Yan
    IEEE ACCESS, 2021, 9 : 67696 - 67705
  • [40] Lightweight Monocular Depth Estimation via Token-Sharing Transformer
    Lee, Dong-Jae
    Lee, Jae Young
    Shon, Hyunguk
    Yi, Eojindl
    Park, Yeong-Hun
    Cho, Sung-Sik
    Kim, Junmo
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 4895 - 4901