Nested block self-attention multiple resolution residual network for multiorgan segmentation from CT

被引:17
|
作者
Jiang, Jue [1 ]
Elguindi, Sharif [1 ]
Berry, Sean L. [1 ]
Onochie, Ifeanyirochukwu [1 ]
Cervino, Laura [1 ]
Deasy, Joseph O. [1 ]
Veeraraghavan, Harini [1 ]
机构
[1] Mem Sloan Kettering Canc Ctr, Dept Med Phys, New York, NY 10065 USA
关键词
abdomen; head and neck; multiple organs CT segmentation; nested-block self-attention; AUTO-SEGMENTATION; HEAD; IMAGES;
D O I
10.1002/mp.15765
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Background Fast and accurate multiorgans segmentation from computed tomography (CT) scans is essential for radiation treatment planning. Self-attention(SA)-based deep learning methodologies provide higher accuracies than standard methods but require memory and computationally intensive calculations, which restricts their use to relatively shallow networks. Purpose Our goal was to develop and test a new computationally fast and memory-efficient bidirectional SA method called nested block self-attention (NBSA), which is applicable to shallow and deep multiorgan segmentation networks. Methods A new multiorgan segmentation method combining a deep multiple resolution residual network with computationally efficient SA called nested block SA (MRRN-NBSA) was developed and evaluated to segment 18 different organs from head and neck (HN) and abdomen organs. MRRN-NBSA combines features from multiple image resolutions and feature levels with SA to extract organ-specific contextual features. Computational efficiency is achieved by using memory blocks of fixed spatial extent for SA calculation combined with bidirectional attention flow. Separate models were trained for HN (n = 238) and abdomen (n = 30) and tested on set aside open-source grand challenge data sets for HN (n = 10) using a public domain database of computational anatomy and blinded testing on 20 cases from Beyond the Cranial Vault data set with overall accuracy provided by the grand challenge website for abdominal organs. Robustness to two-rater segmentations was also evaluated for HN cases using the open-source data set. Statistical comparison of MRRN-NBSA against Unet, convolutional network-based SA using criss-cross attention (CCA), dual SA, and transformer-based (UNETR) methods was done by measuring the differences in the average Dice similarity coefficient (DSC) accuracy for all HN organs using the Kruskall-Wallis test, followed by individual method comparisons using paired, two-sided Wilcoxon-signed rank tests at 95% confidence level with Bonferroni correction used for multiple comparisons. Results MRRN-NBSA produced an average high DSC of 0.88 for HN and 0.86 for the abdomen that exceeded current methods. MRRN-NBSA was more accurate than the computationally most efficient CCA (average DSC of 0.845 for HN, 0.727 for abdomen). Kruskal-Wallis test showed significant difference between evaluated methods (p=0.00025). Pair-wise comparisons showed significant differences between MRRN-NBSA than Unet (p=0.0003), CCA (p=0.030), dual (p=0.038), and UNETR methods (p=0.012) after Bonferroni correction. MRRN-NBSA produced less variable segmentations for submandibular glands (0.82 +/- 0.06) compared to two raters (0.75 +/- 0.31). Conclusions MRRN-NBSA produced more accurate multiorgan segmentations than current methods on two different public data sets. Testing on larger institutional cohorts is required to establish feasibility for clinical use.
引用
收藏
页码:5244 / 5257
页数:14
相关论文
共 50 条
  • [1] A hybrid network for three-dimensional seismic fault segmentation based on nested residual attention and self-attention mechanism
    Sun, Qifeng
    Jiang, Hui
    Du, Qizhen
    Gong, Faming
    GEOPHYSICAL PROSPECTING, 2025, 73 (02) : 575 - 594
  • [2] Multiple Self-attention Network for Intracranial Vessel Segmentation
    Li, Yang
    Ni, Jiajia
    Elazab, Ahmed
    Wu, Jianhuang
    2021 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2021,
  • [3] Face Super-Resolution Reconstruction Based on Self-Attention Residual Network
    Liu, Qing-Ming
    Jia, Rui-Sheng
    Zhao, Chao-Yue
    Liu, Xiao-Ying
    Sun, Hong-Mei
    Zhang, Xing-Li
    IEEE ACCESS, 2020, 8 : 4110 - 4121
  • [4] Lightweight Self-Attention Network for Semantic Segmentation
    Zhou, Yan
    Zhou, Haibin
    Li, Nanjun
    Li, Jianxun
    Wang, Dongli
    2022 INTERNATIONAL JOINT CONFERENCE ON NEURAL NETWORKS (IJCNN), 2022,
  • [5] A Dynamic Residual Self-Attention Network for Lightweight Single Image Super-Resolution
    Park, Karam
    Soh, Jae Woong
    Cho, Nam Ik
    IEEE TRANSACTIONS ON MULTIMEDIA, 2023, 25 : 907 - 918
  • [6] Multilayer self-attention residual network for code search
    Hu, Haize
    Liu, Jianxun
    Zhang, Xiangping
    CONCURRENCY AND COMPUTATION-PRACTICE & EXPERIENCE, 2023, 35 (09):
  • [7] Lightweight Self-Attention Residual Network for Hyperspectral Classification
    Xia, Jinbiao
    Cui, Ying
    Li, Wenshan
    Wang, Liguo
    Wang, Chao
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2022, 19
  • [8] Self-attention feature fusion network for semantic segmentation
    Zhou, Zhen
    Zhou, Yan
    Wang, Dongli
    Mu, Jinzhen
    Zhou, Haibin
    NEUROCOMPUTING, 2021, 453 : 50 - 59
  • [9] Investigating Self-Attention Network for Chinese Word Segmentation
    Gan, Leilei
    Zhang, Yue
    IEEE-ACM TRANSACTIONS ON AUDIO SPEECH AND LANGUAGE PROCESSING, 2020, 28 : 2933 - 2941
  • [10] SN-FPN: Self-Attention Nested Feature Pyramid Network for Digital Pathology Image Segmentation
    Lee, Sanghoon
    Aminul Islam, Kazi
    Chandana Koganti, Sai
    Yaganti, Varshini
    Ramya Sri Mamillapalli, Sai
    Vitalos, Hannah
    Williamson, Drew F. K.
    IEEE ACCESS, 2024, 12 : 92764 - 92773