Two-View Correspondence Learning With Local Consensus Transformer

被引:0
|
作者
Wang, Gang [1 ]
Chen, Yufei [2 ]
机构
[1] Shanghai Univ Finance & Econ, Sch Stat & Management, Shanghai 200433, Peoples R China
[2] Tongji Univ, Sch Comp Sci & Technol, Shanghai 201804, Peoples R China
基金
中国国家自然科学基金;
关键词
Feature extraction; Transformers; Robustness; Technological innovation; Network architecture; Image reconstruction; Image edge detection; Geometry; Computer vision; Computer architecture; Correspondence learning; feature matching; local consensus (LC); transformer;
D O I
10.1109/TNNLS.2024.3488197
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Correspondence learning is a crucial component in multiview geometry and computer vision. The presence of heavy outliers (mismatches) consistently renders the matching problem to be highly challenging. In this article, we revisit the benefits of local consensus (LC) in traditional feature matching and introduce the concept of LC to design a trainable neural network capable of capturing the underlying correspondences. This network is named the LC transformer (LCT) and is specifically tailored for wide-baseline stereo applications. Our network architecture comprises three distinct operations. To establish the neighbor topology, we employ a dynamic graph-based embedding layer as the initial step. Subsequently, these local topologies serve as guidance for the multihead self-attention layer, enabling it to extract a more extensive contextual understanding through channel attention (CA). Following this, order-aware graph pooling is applied to extract the global context information from the embedded LC. Through the experimental analysis, the ablation study reveals that PointNet-like learning models can, indeed, benefit from the incorporation of LC. The proposed model achieves state-of-the-art performance in both challenging scenes, namely, the YFCC100M outdoor and SUN3D indoor environments, even in the presence of more than 90% outliers.
引用
收藏
页数:14
相关论文
共 50 条
  • [31] NCMNet: Neighbor Consistency Mining Network for Two-View Correspondence Pruning
    Liu, Xin
    Qin, Rong
    Yan, Junchi
    Yang, Jufeng
    IEEE TRANSACTIONS ON PATTERN ANALYSIS AND MACHINE INTELLIGENCE, 2024, 46 (12) : 11254 - 11272
  • [32] MC-Net: Integrating Multi-Level Geometric Context for Two-View Correspondence Learning
    Li, Zizhuo
    Su, Chunbao
    Fan, Fan
    Huang, Jun
    Ma, Jiayi
    IEEE TRANSACTIONS ON CIRCUITS AND SYSTEMS FOR VIDEO TECHNOLOGY, 2024, 34 (08) : 7550 - 7565
  • [33] PMA-Net: Progressive multi-stage adaptive feature learning for two-view correspondence
    Li, Xiaojie
    Zhuang, Fengyuan
    Liu, Yizhang
    Chen, Riqing
    Wei, Lifang
    Yang, Changcai
    KNOWLEDGE-BASED SYSTEMS, 2024, 298
  • [34] Two-view tomography
    Denisova, NV
    JOURNAL OF PHYSICS D-APPLIED PHYSICS, 2000, 33 (03) : 313 - 319
  • [35] When Transformer Meets Large Graphs: An Expressive and Efficient Two-View Architecture
    Kuang, Weirui
    Wang, Zhen
    Wei, Zhewei
    Li, Yaliang
    Ding, Bolin
    IEEE TRANSACTIONS ON KNOWLEDGE AND DATA ENGINEERING, 2024, 36 (10) : 5440 - 5452
  • [36] Uncalibrated two-view metrology
    Liang, BJ
    Chen, ZZ
    Pears, N
    PROCEEDINGS OF THE 17TH INTERNATIONAL CONFERENCE ON PATTERN RECOGNITION, VOL 1, 2004, : 96 - 99
  • [37] Two-view screening in UK
    不详
    EUROPEAN JOURNAL OF CANCER, 2003, 39 (04) : 410 - 410
  • [38] Two-view Matching with View Synthesis Revisited
    Mishkin, Dmytro
    Perdoch, Michal
    Matas, Jiri
    PROCEEDINGS OF 2013 28TH INTERNATIONAL CONFERENCE ON IMAGE AND VISION COMPUTING NEW ZEALAND (IVCNZ 2013), 2013, : 436 - 441
  • [39] Dual Memory Neural Computerfor Asynchronous Two-view Sequential Learning
    Hung Le
    Truyen Tran
    Venkatesh, Svetha
    KDD'18: PROCEEDINGS OF THE 24TH ACM SIGKDD INTERNATIONAL CONFERENCE ON KNOWLEDGE DISCOVERY & DATA MINING, 2018, : 1637 - 1645
  • [40] Scalable variable selection for two-view learning tasks with projection operators
    Szedmak, Sandor
    Huusari, Riikka
    Le, Tat Hong Duong
    Rousu, Juho
    MACHINE LEARNING, 2024, 113 (06) : 3525 - 3544