Two-View Correspondence Learning With Local Consensus Transformer

被引:0
|
作者
Wang, Gang [1 ]
Chen, Yufei [2 ]
机构
[1] Shanghai Univ Finance & Econ, Sch Stat & Management, Shanghai 200433, Peoples R China
[2] Tongji Univ, Sch Comp Sci & Technol, Shanghai 201804, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Transformers; Robustness; Technological innovation; Network architecture; Image reconstruction; Image edge detection; Geometry; Computer vision; Computer architecture; Correspondence learning; feature matching; local consensus (LC); transformer;
D O I
10.1109/TNNLS.2024.3488197
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Correspondence learning is a crucial component in multiview geometry and computer vision. The presence of heavy outliers (mismatches) consistently renders the matching problem to be highly challenging. In this article, we revisit the benefits of local consensus (LC) in traditional feature matching and introduce the concept of LC to design a trainable neural network capable of capturing the underlying correspondences. This network is named the LC transformer (LCT) and is specifically tailored for wide-baseline stereo applications. Our network architecture comprises three distinct operations. To establish the neighbor topology, we employ a dynamic graph-based embedding layer as the initial step. Subsequently, these local topologies serve as guidance for the multihead self-attention layer, enabling it to extract a more extensive contextual understanding through channel attention (CA). Following this, order-aware graph pooling is applied to extract the global context information from the embedded LC. Through the experimental analysis, the ablation study reveals that PointNet-like learning models can, indeed, benefit from the incorporation of LC. The proposed model achieves state-of-the-art performance in both challenging scenes, namely, the YFCC100M outdoor and SUN3D indoor environments, even in the presence of more than 90% outliers.
引用
收藏
页数:14
相关论文
共 50 条
  • [1] Local Consensus Enhanced Siamese Network with Reciprocal Loss for Two-view Correspondence Learning
    Wang, Linbo
    Wu, Jing
    Fang, Xianyong
    Liu, Zhengyi
    Cao, Chenjie
    Fu, Yanwei
    PROCEEDINGS OF THE 31ST ACM INTERNATIONAL CONFERENCE ON MULTIMEDIA, MM 2023, 2023, : 5235 - 5243
  • [2] VRTNet: Vector Rectifier Transformer for Two-View Correspondence Learning
    Yang, Meng
    Chen, Jun
    Tian, Xin
    Wei, Longsheng
    Ma, Jiayi
    IEEE TRANSACTIONS ON MULTIMEDIA, 2025, 27 : 515 - 530
  • [3] BCLNet: Bilateral Consensus Learning for Two-View Correspondence Pruning
    Miao, Xiangyang
    Xiao, Guobao
    Wang, Shiping
    Yu, Jun
    THIRTY-EIGHTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 38 NO 5, 2024, : 4225 - 4232
  • [4] Correspondence Attention Transformer: A Context-Sensitive Network for Two-View Correspondence Learning
    Ma, Jiayi
    Wang, Yang
    Fan, Aoxiang
    Xiao, Guobao
    Chen, Riqing
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 3509 - 3524
  • [5] Local Consensus Transformer for Correspondence Learning
    Wang, Gang
    Chen, Yufei
    2023 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO, ICME, 2023, : 1151 - 1156
  • [6] Core sample consensus method for two-view correspondence matching
    Ding, Xintao
    Li, Boquan
    Zhou, Wen
    Zhao, Cheng
    MULTIMEDIA TOOLS AND APPLICATIONS, 2023, 83 (8) : 24609 - 24630
  • [7] Core sample consensus method for two-view correspondence matching
    Xintao Ding
    Boquan Li
    Wen Zhou
    Cheng Zhao
    Multimedia Tools and Applications, 2024, 83 : 24609 - 24630
  • [8] Correction to: Core sample consensus method for two-view correspondence matching
    Xintao Ding
    Boquan Li
    Wen Zhou
    Cheng Zhao
    Multimedia Tools and Applications, 2024, 83 (8) : 24631 - 24631
  • [9] Enhancing two-view correspondence learning by local-global self-attention
    Dai, Luanyuan
    Liu, Xin
    Liu, Yizhang
    Yang, Changcai
    Wei, Lifang
    Lin, Yaohai
    Chen, Riqing
    NEUROCOMPUTING, 2021, 459 : 176 - 187
  • [10] ConvMatch: Rethinking Network Design for Two-View Correspondence Learning
    Zhang, Shihua
    Ma, Jiayi
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 3, 2023, : 3472 - 3479