Structured Attention Guided Convolutional Neural Fields for Monocular Depth Estimation

被引:250
|
作者
Xu, Dan [1 ]
Wang, Wei [1 ]
Tang, Hao [1 ]
Liu, Hong [2 ]
Sebe, Nicu [1 ]
Ricci, Elisa [1 ,3 ]
机构
[1] Univ Trento, Multimedia & Human Understanding Grp, Trento, Italy
[2] Peking Univ, Shenzhen Grad Sch, Key Lab Machine Percept, Beijing, Peoples R China
[3] Fdn Bruno Kessler, Technol Vis Grp, Trento, Italy
基金
中国国家自然科学基金;
关键词
D O I
10.1109/CVPR.2018.00412
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Recent works have shown the benefit of integrating Conditional Random Fields (CRFs) models into deep architectures for improving pixel-level prediction tasks. Following this line of research, in this paper we introduce a novel approach for monocular depth estimation. Similarly to previous works, our method employs a continuous CRF to fuse multi-scale information derived from different layers of a front-end Convolutional Neural Network (CNN). Differently from past works, our approach benefits from a structured attention model which automatically regulates the amount of information transferred between corresponding features at different scales. Importantly, the proposed attention model is seamlessly integrated into the CRF allowing end-to-end training of the entire architecture. Our extensive experimental evaluation demonstrates the effectiveness of the proposed method which is competitive with previous methods on the KITH benchmark and outperforms the state of the art on the NYU Depth V2 dataset.
引用
收藏
页码:3917 / 3925
页数:9
相关论文
共 50 条
  • [41] Accurate depth estimation in structured light fields
    Cai, Zewei
    Liu, Xiaoli
    Pedrini, Giancarlo
    Osten, Wolfgang
    Peng, Xiang
    OPTICS EXPRESS, 2019, 27 (09) : 13532 - 13546
  • [42] Patch-Wise Attention Network for Monocular Depth Estimation
    Lee, Sihaeng
    Lee, Janghyeon
    Kim, Byungju
    Yi, Eojindl
    Kim, Junmo
    THIRTY-FIFTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, THIRTY-THIRD CONFERENCE ON INNOVATIVE APPLICATIONS OF ARTIFICIAL INTELLIGENCE AND THE ELEVENTH SYMPOSIUM ON EDUCATIONAL ADVANCES IN ARTIFICIAL INTELLIGENCE, 2021, 35 : 1873 - 1881
  • [43] Radar Fusion Monocular Depth Estimation Based on Dual Attention
    Long, JianYu
    Huang, JinGui
    Wang, ShengChun
    ARTIFICIAL INTELLIGENCE AND SECURITY, ICAIS 2022, PT I, 2022, 13338 : 166 - 179
  • [44] DEEP MONOCULAR VIDEO DEPTH ESTIMATION USING TEMPORAL ATTENTION
    Ren, Haoyu
    El-khamy, Mostafa
    Lee, Jungwon
    2020 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH, AND SIGNAL PROCESSING, 2020, : 1988 - 1992
  • [45] DAttNet: monocular depth estimation network based on attention mechanisms
    Astudillo, Armando
    Barrera, Alejandro
    Guindel, Carlos
    Al-Kaff, Abdulla
    Garcia, Fernando
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (07): : 3347 - 3356
  • [46] Monocular Depth Estimation with Optical Flow Attention for Autonomous Drones
    Shimhada, Tomoyasu
    Nishikawa, Hiroki
    Kong, Xiangbo
    Tomiyama, Hiroyuki
    2022 19TH INTERNATIONAL SOC DESIGN CONFERENCE (ISOCC), 2022, : 197 - 198
  • [47] FG-Depth: Flow-Guided Unsupervised Monocular Depth Estimation
    Zhu, Junyu
    Liu, Lina
    Liu, Yong
    Li, Wanlong
    Wen, Feng
    Zhang, Hongbo
    2023 IEEE INTERNATIONAL CONFERENCE ON ROBOTICS AND AUTOMATION, ICRA, 2023, : 4924 - 4930
  • [48] Monocular depth estimation with multi-view attention autoencoder
    Jung, Geunho
    Yoon, Sang Min
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (23) : 33759 - 33770
  • [49] Boosting Monocular Depth Estimation with Channel Attention and Mutual Learning
    Takagi, Kazunari
    Ito, Seiya
    Kaneko, Naoshi
    Sumi, Kazuhiko
    2019 JOINT 8TH INTERNATIONAL CONFERENCE ON INFORMATICS, ELECTRONICS & VISION (ICIEV) AND 2019 3RD INTERNATIONAL CONFERENCE ON IMAGING, VISION & PATTERN RECOGNITION (ICIVPR) WITH INTERNATIONAL CONFERENCE ON ACTIVITY AND BEHAVIOR COMPUTING (ABC), 2019, : 228 - 233
  • [50] MAMo: Leveraging Memory and Attention for Monocular Video Depth Estimation
    Yasarla, Rajeev
    Cai, Hong
    Jeong, Jisoo
    Shi, Yunxiao
    Garrepalli, Risheek
    Porikli, Fatih
    2023 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2023), 2023, : 8720 - 8730