DAE-Former: Dual Attention-Guided Efficient Transformer for Medical Image Segmentation

被引:48
|
作者
Azad, Reza [1 ]
Arimond, Rene [1 ]
Aghdam, Ehsan Khodapanah [2 ]
Kazerouni, Amirhossein [3 ]
Merhof, Dorit [4 ,5 ]
机构
[1] Rhein Westfal TH Aachen, Fac Elect Engn & Informat Technol, Aachen, Germany
[2] Shahid Beheshti Univ, Dept Elect Engn, Tehran, Iran
[3] Iran Univ Sci & Technol, Sch Elect Engn, Tehran, Iran
[4] Univ Regensburg, Inst Image Anal & Comp Vis, Fac Informat & Data Sci, Regensburg, Germany
[5] Fraunhofer Inst Digital Med MEVIS, Bremen, Germany
关键词
Transformer; Segmentation; Deep Learning; Medical;
D O I
10.1007/978-3-031-46005-0_8
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Transformers have recently gained attention in the computer vision domain due to their ability to model long-range dependencies. However, the self-attention mechanism, which is the core part of the Transformer model, usually suffers from quadratic computational complexity with respect to the number of tokens. Many architectures attempt to reduce model complexity by limiting the self-attention mechanism to local regions or by redesigning the tokenization process. In this paper, we propose DAE-Former, a novel method that seeks to provide an alternative perspective by efficiently designing the self-attention mechanism. More specifically, we reformulate the self-attention mechanism to capture both spatial and channel relations across the whole feature dimension while staying computationally efficient. Furthermore, we redesign the skip connection path by including the cross-attention module to ensure the feature reusability and enhance the localization power. Our method outperforms state-of-the-art methods on multi-organ cardiac and skin lesion segmentation datasets, without pre-training weights. The code is publicly available at GitHub.
引用
收藏
页码:83 / 95
页数:13
相关论文
共 50 条
  • [31] Unsupervised Attention-guided Image-to-Image Translation
    Mejjati, Youssef A.
    Richardt, Christian
    Tompkin, James
    Cosker, Darren
    Kim, Kwang In
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 31 (NIPS 2018), 2018, 31
  • [32] Hierarchical volumetric transformer with comprehensive attention for medical image segmentation
    Zhang, Zhuang
    Luo, Wenjie
    MATHEMATICAL BIOSCIENCES AND ENGINEERING, 2023, 20 (02) : 3177 - 3190
  • [33] ConTrans: Improving Transformer with Convolutional Attention for Medical Image Segmentation
    Lin, Ailiang
    Xu, Jiayu
    Li, Jinxing
    Lu, Guangming
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2022, PT V, 2022, 13435 : 297 - 307
  • [34] Dual Attention-Guided Detail and Structure Information Fusion Network for Image Dehazing
    Gao J.-R.
    Li H.-F.
    Zhang Y.-F.
    Xie M.-H.
    Li F.
    Tien Tzu Hsueh Pao/Acta Electronica Sinica, 2023, 51 (01): : 160 - 171
  • [35] Few Shot Medical Image Segmentation with Cross Attention Transformer
    Lin, Yi
    Chen, Yufan
    Cheng, Kwang-Ting
    Chen, Hao
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT II, 2023, 14221 : 233 - 243
  • [36] Attention-Guided Multispectral and Panchromatic Image Classification
    Shi, Cheng
    Dang, Yenan
    Fang, Li
    Lv, Zhiyong
    Shen, Huifang
    REMOTE SENSING, 2021, 13 (23)
  • [37] Dual Cross-Attention for medical image segmentation
    Ates, Gorkem Can
    Mohan, Prasoon
    Celik, Emrah
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2023, 126
  • [38] Attention-Guided Memory Model for Video Object Segmentation
    Lin, Yunjian
    Tan, Yihua
    Communications in Computer and Information Science, 2022, 1566 CCIS : 67 - 85
  • [39] Attention-guided chained context aggregation for semantic segmentation*
    Tang, Quan
    Liu, Fagui
    Zhang, Tong
    Jiang, Jun
    Zhang, Yu
    IMAGE AND VISION COMPUTING, 2021, 115 (115)
  • [40] Attention-guided Adversarial Attack for Video Object Segmentation
    Yao, Rui
    Chen, Ying
    Zhou, Yong
    Hu, Fuyuan
    Zhao, Jiaqi
    Liu, Bing
    Shao, Zhiwen
    ACM TRANSACTIONS ON INTELLIGENT SYSTEMS AND TECHNOLOGY, 2023, 14 (06)