Multi-view Instance Attention Fusion Network for classification

被引:5
|
作者
Li, Jinxing [1 ,2 ]
Zhou, Chuhao [1 ]
Ji, Xiaoqiang [3 ,4 ]
Li, Mu [1 ]
Lu, Guangming [1 ]
Xu, Yong [1 ,2 ]
Zhang, David [3 ]
机构
[1] Harbin Inst Technol, Shenzhen, Peoples R China
[2] Shenzhen Key Lab Visual Object Detect & Recognit, Shenzhen, Peoples R China
[3] Chinese Univ Hong Kong, Shenzhen, Peoples R China
[4] Shenzhen Inst Artificial Intelligence & Robot Soc, Shenzhen, Peoples R China
关键词
Multi-view; Instance learning; Classification; Cross-fusion; RECOGNITION; NEAREST;
D O I
10.1016/j.inffus.2023.101974
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Multi-view learning for classification has achieved a remarkable performance compared with the single-view based methods. Inspired by the instance based learning which directly regards the instance as the prior and well preserves the valuable information in different instances, a Multi-view Instance Attention Fusion Network (MvIAFN) is proposed to efficiently exploit the correlation across both instances and views. Specifically, a small number of instances from different views are first sampled as the set of templates. Given an additional instance and based on the similarities between it and the selected templates, it can be re-presented by following an attention strategy. Thanks for this strategy, the given instance is capable of preserving the additional information from the selected instances, achieving the purpose of extracting the instance-correlation. Additionally, for each sample, we not only perform the instance attention in each single view but also get the attention across multiple views, allowing us to further fuse them to obtain the fused attention for each view. Experimental results on datasets substantiate the effectiveness of our proposed method compared with state-of-the-arts.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] Sequential attention layer-wise fusion network for multi-view classification
    Teng, Qing
    Yang, Xibei
    Sun, Qiguo
    Wang, Pingxin
    Wang, Xun
    Xu, Taihua
    INTERNATIONAL JOURNAL OF MACHINE LEARNING AND CYBERNETICS, 2024, 15 (12) : 5549 - 5561
  • [2] Fusion of heterogeneous attention mechanisms in multi-view convolutional neural network for text classification
    Liang, Yunji
    Li, Huihui
    Guo, Bin
    Yu, Zhiwen
    Zheng, Xiaolong
    Samtani, Sagar
    Zeng, Daniel D.
    INFORMATION SCIENCES, 2021, 548 : 295 - 312
  • [3] GAF-Net: Graph attention fusion network for multi-view semi-supervised classification
    Song, Na
    Du, Shide
    Wu, Zhihao
    Zhong, Luying
    Yang, Laurence T.
    Yang, Jing
    Wang, Shiping
    EXPERT SYSTEMS WITH APPLICATIONS, 2024, 238
  • [4] MVFFNet: Multi-view feature fusion network for imbalanced ship classification
    Liang, Maohan
    Zhan, Yang
    Liu, Ryan Wen
    PATTERN RECOGNITION LETTERS, 2021, 151 : 26 - 32
  • [5] Image classification with multi-view multi-instance metric learning
    Tang, Jingjing
    Li, Dewei
    Tian, Yingjie
    EXPERT SYSTEMS WITH APPLICATIONS, 2022, 189
  • [6] MDF-Net: A multi-view dual-attention fusion network for efficient bird sound classification
    Xie, Shanshan
    Xie, Jiangjian
    Zhang, Junguo
    Zhang, Yan
    Wang, Lifeng
    Hu, Huijian
    APPLIED ACOUSTICS, 2024, 225
  • [7] Multi-view features fusion for birdsong classification
    Xie, Shanshan
    Lu, Jing
    Liu, Jiang
    Zhang, Yan
    Lv, Danjv
    Chen, Xu
    Zhao, Youjie
    ECOLOGICAL INFORMATICS, 2022, 72
  • [8] Multi-view Feature Fusion for Activity Classification
    Hekmat, Mitra
    Mousavi, Zahra
    Aghajan, Hamid
    ICDSC 2016: 10TH INTERNATIONAL CONFERENCE ON DISTRIBUTED SMART CAMERA, 2016, : 190 - 195
  • [9] Multi-view stereo network with point attention
    Zhao, Rong
    Gu, Zhuoer
    Han, Xie
    He, Ligang
    Sun, Fusheng
    Jiao, Shichao
    APPLIED INTELLIGENCE, 2023, 53 (22) : 26622 - 26636
  • [10] Multi-view stereo network with point attention
    Rong Zhao
    Zhuoer Gu
    Xie Han
    Ligang He
    Fusheng Sun
    Shichao Jiao
    Applied Intelligence, 2023, 53 : 26622 - 26636