PSP-MVSNet: Deep Patch-Based Similarity Perceptual for Multi-view Stereo Depth Inference

被引:2
|
作者
Jie, Leiping [1 ,2 ]
Zhang, Hui [2 ]
机构
[1] Hong Kong Baptist Univ, Dept Comp Sci, Hong Kong, Peoples R China
[2] BNU HKBU United Int Coll, Guangdong Key Lab Interdisciplinary Res & Applica, Zhuhai, Peoples R China
基金
中国国家自然科学基金;
关键词
Depth estimation; Patch-based similarity; Dynamic depth range; Multi-view stereo;
D O I
10.1007/978-3-031-15919-0_27
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
This paper proposes PSP-MVSNet for depth inference problem in multi-view stereo (MVS). We first introduce a novel patch-based similarity perceptual (PSP) module for effectively constructing 3D cost volume. Unlike previous methods that leverage variance-based operators to fuse feature volumes of different views, our method leverages a cosine similarity measure to calculate matching scores for pairs of deep feature vectors and then treats these scores as weights for constructing the 3D cost volume. This is based on an important observation that many performance degradation factors, e.g., illumination changes or occlusions, will lead to pixel differences between multi-view images. We demonstrate that a patch-based cosine similarity can be used as explicit supervision for feature learning and can help speed up convergence. Furthermore, To adaptively set different depth ranges for different pixels, we extend an existing dynamic depth range searching method with a simple yet effective improvement. We can use this improved searching method to train our model in an end-to-end manner and further improve the performance of our method. Experimental results show that our method achieves state-of-the-art performance on the DTU dataset and comparative results on the intermediate set of Tanks and Temples dataset.
引用
收藏
页码:316 / 328
页数:13
相关论文
共 50 条
  • [41] Depth Map Based Facade Abstraction from Noisy Multi-View Stereo Point Clouds
    Ley, Andreas
    Hellwich, Olaf
    PATTERN RECOGNITION, GCPR 2016, 2016, 9796 : 155 - 165
  • [42] Comparative Analysis of Deep Learning-Based Stereo Matching and Multi-View Stereo for Urban DSM Generation
    Reyes, Mario Fuentes
    d'Angelo, Pablo
    Fraundorfer, Friedrich
    REMOTE SENSING, 2025, 17 (01)
  • [43] SA-MVSNet: Self-attention-based multi-view stereo network for 3D reconstruction of images with weak texture
    Yang, Ronghao
    Miao, Wang
    Zhang, Zhenxin
    Liu, Zhenlong
    Li, Mubai
    Lin, Bin
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 131
  • [44] A Confidence-based Iterative Solver of Depths and Surface Normals for Deep Multi-view Stereo
    Zhao, Wang
    Liu, Shaohui
    Wei, Yi
    Guo, Hengkai
    Liu, Yong-Jin
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 6148 - 6157
  • [45] Instructed fine-tuning based on semantic consistency constraint for deep multi-view stereo
    Zhang, Yan
    Yan, Hongping
    Ding, Kun
    Cai, Tingting
    Zhou, Yueyue
    APPLIED INTELLIGENCE, 2025, 55 (06)
  • [46] Graph Representation-Based Deep Multi-View Semantic Similarity Learning Model for Recommendation
    Song, Jiagang
    Song, Jiayu
    Yuan, Xinpan
    He, Xiao
    Zhu, Xinghui
    FUTURE INTERNET, 2022, 14 (02)
  • [47] Deployment of a deep-learning based multi-view stereo approach for measurement of ship shell plates
    He, Pengpeng
    Hu, Delin
    Hu, Yong
    OCEAN ENGINEERING, 2022, 260
  • [48] Prior depth-based multi-view stereo network for online 3D model reconstruction
    Song, Soohwan
    Truong, Khang Giang
    Kim, Daekyum
    Jo, Sungho
    PATTERN RECOGNITION, 2023, 136
  • [49] Rational Polynomial Camera Model Warping for Deep Learning Based Satellite Multi-View Stereo Matching
    Gao, Jian
    Liu, Jin
    Ji, Shunping
    2021 IEEE/CVF INTERNATIONAL CONFERENCE ON COMPUTER VISION (ICCV 2021), 2021, : 6128 - 6137
  • [50] Deep learning based multi-view dense matching with joint depth and surface normal estimation
    Liu, Jin
    Ji, Shunping
    Cehui Xuebao/Acta Geodaetica et Cartographica Sinica, 2025, 53 (12): : 2391 - 2403