COVID-19 diagnosis based on swin transformer model with demographic information fusion and enhanced multi-head attention mechanism

被引:3
|
作者
Sun, Yunlong [1 ]
Lian, Jingge [2 ,3 ,4 ]
Teng, Ze [5 ]
Wei, Ziyi [3 ]
Tang, Yi [3 ]
Yang, Liu [3 ]
Gao, Yajuan [2 ,6 ]
Wang, Tianfu [1 ]
Li, Hongfeng [3 ]
Xu, Meng [3 ,4 ]
Lei, Baiying [1 ]
机构
[1] Shenzhen Univ, Med Sch, Sch Biomed Engn, Natl Reg Key Technol Engn Lab Med Ultrasound,Guang, Nanhai Ave 3688, Shenzhen, Guangdong, Peoples R China
[2] Peking Univ Third Hosp, Dept Radiol, Beijing, Peoples R China
[3] Peking Univ, Inst Med Technol, Hlth Sci Ctr, Beijing, Peoples R China
[4] Beijing Key Lab Magnet Resonance Imaging Device &, Beijing, Peoples R China
[5] Chinese Acad Med Sci & Peking Union Med Coll, Natl Clin Res Ctr Canc, Natl Canc Ctr,Dept Radiol,Canc Hosp, Beijing 100021, Peoples R China
[6] NMPA Key Lab Evaluat Med Imaging Equipment & Tech, Beijing, Peoples R China
关键词
COVID-19; diagnosis; Swin Transformer; Demographic information fusion; Enhanced Multi-head Self-Attention; SYSTEM;
D O I
10.1016/j.eswa.2023.122805
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Coronavirus disease 2019 (COVID-19) is an acute disease, which can rapidly become severe. Hence, it is of great significance to realize the automatic diagnosis of COVID-19. However, existing models are often inapplicable for fusing patients' demographic information due to its low dimensionality. To address this, we propose a COVID-19 patient diagnosis method with feature fusion and a model based on Swin Transformer. Specifically, two auxiliary tasks are added for fusing computed tomography (CT) images and patients' demographic information, which utilizes the patients' demographic information as the label for the auxiliary tasks. Besides, our approach involves designing a Swin Transformer model with Enhanced Multi-head Self-Attention (EMSA) to capture different features from CT data. Meanwhile, the EMSA module is able to extract and fuse attention information in different representation subspaces, further enhancing the performance of the model. Furthermore, we evaluate our model in COVIDx CT-3 dataset with different tasks to classify Normal Controls (NC), COVID-19 cases and community-acquired pneumonia (CAP) cases and compare the performance of our method with other models, which show the effectiveness of our model. In addition, we have conducted various visualization efforts to demonstrate the interpretability of our model, including principal component analysis, attention heatmaps, etc. Various results indicate that our model is capable of making reasonable diagnosis.
引用
收藏
页数:11
相关论文
共 50 条
  • [1] COVID-19 Diagnosis Based on Swin Transformer Model with Demographic Information Fusion and Enhanced Multi-head Attention Mechanism
    Sun, Yunlong
    Liu, Yiyao
    Qu, Junlong
    Dong, Xiang
    Song, Xuegang
    Lei, Baiying
    MACHINE LEARNING IN MEDICAL IMAGING, MLMI 2023, PT II, 2024, 14349 : 194 - 204
  • [2] Diversifying Multi-Head Attention in the Transformer Model
    Ampazis, Nicholas
    Sakketou, Flora
    MACHINE LEARNING AND KNOWLEDGE EXTRACTION, 2024, 6 (04): : 2618 - 2638
  • [3] MSKd_Net: Multi-Head Attention-Based Swin Transformer for Kidney Diseases Classification
    Sharen, H.
    Narendra, Modigari
    Anbarasi, L. Jani
    IEEE ACCESS, 2024, 12 : 181975 - 181986
  • [4] A multi-scale gated multi-head attention depthwise separable CNN model for recognizing COVID-19
    Geng Hong
    Xiaoyan Chen
    Jianyong Chen
    Miao Zhang
    Yumeng Ren
    Xinyu Zhang
    Scientific Reports, 11
  • [5] Network Configuration Entity Extraction Method Based on Transformer with Multi-Head Attention Mechanism
    Yang, Yang
    Qu, Zhenying
    Yan, Zefan
    Gao, Zhipeng
    Wang, Ti
    CMC-COMPUTERS MATERIALS & CONTINUA, 2024, 78 (01): : 735 - 757
  • [6] A multi-scale gated multi-head attention depthwise separable CNN model for recognizing COVID-19
    Hong, Geng
    Chen, Xiaoyan
    Chen, Jianyong
    Zhang, Miao
    Ren, Yumeng
    Zhang, Xinyu
    SCIENTIFIC REPORTS, 2021, 11 (01)
  • [7] Machine Reading Comprehension Model Based on Multi-head Attention Mechanism
    Xue, Yong
    ADVANCED INTELLIGENT TECHNOLOGIES FOR INDUSTRY, 2022, 285 : 45 - 58
  • [8] Diagnostic Model for Transformer Core Loosening Faults Based on the Gram Angle Field and Multi-Head Attention Mechanism
    Chen, Junyu
    Duan, Nana
    Zhou, Xikun
    Wang, Ziyu
    APPLIED SCIENCES-BASEL, 2024, 14 (23):
  • [9] A Network Intrusion Detection Model Based on BiLSTM with Multi-Head Attention Mechanism
    Zhang, Jingqi
    Zhang, Xin
    Liu, Zhaojun
    Fu, Fa
    Jiao, Yihan
    Xu, Fei
    ELECTRONICS, 2023, 12 (19)
  • [10] Feature Fusion and Multi-head Attention Based Hindi Captioner
    Meghwal, Virendra Kumar
    Mittal, Namita
    Singh, Girdhari
    COMPUTER VISION AND IMAGE PROCESSING, CVIP 2023, PT I, 2024, 2009 : 479 - 487