Image-level supervision and self-training for transformer-based cross-modality tumor segmentation

被引:0
|
作者
d'Assier, Malo Alefsen de Boisredon [1 ]
Portafaix, Aloys [1 ,2 ]
Vorontsov, Eugene [3 ]
Le, William Trung [1 ,2 ]
Kadoury, Samuel [1 ,2 ]
机构
[1] Polytech Montreal, Montreal, PQ, Canada
[2] Univ Montreal, Ctr Rech Ctr Hosp, Montreal, PQ, Canada
[3] Paige, Montreal, PQ, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Tumor Segmentation; Semi-supervised Learning; Domain adaptation; Self-training; UNSUPERVISED DOMAIN ADAPTATION;
D O I
10.1016/j.media.2024.103287
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks are commonly used for automated medical image segmentation, but models will frequently struggle to generalize well across different imaging modalities. This issue is particularly problematic due to the limited availability of annotated data, both in the target as well as the source modality, making it difficult to deploy these models on a larger scale. To overcome these challenges, we propose a new semi- supervised training strategy called MoDATTS. Our approach is designed for accurate cross-modality 3D tumor segmentation on unpaired bi-modal datasets. An image-to-image translation strategy between modalities is used to produce synthetic but annotated images and labels in the desired modality and improve generalization to the unannotated target modality. We also use powerful vision transformer architectures for both image translation (TransUNet) and segmentation (Medformer) tasks and introduce an iterative self-training procedure in the later task to further close the domain gap between modalities, thus also training on unlabeled images in the target modality. MoDATTS additionally allows the possibility to exploit image-level labels with a semi-supervised objective that encourages the model to disentangle tumors from the background. This semi-supervised methodology helps in particular to maintain downstream segmentation performance when pixel-level label scarcity is also present in the source modality dataset, or when the source dataset contains healthy controls. The proposed model achieves superior performance compared to other methods from participating teams in the CrossMoDA 2022 vestibular schwannoma (VS) segmentation challenge, as evidenced by its reported top Dice score of 0.87 +/- 0.04 . 87 +/- 0 . 04 for the VS segmentation. MoDATTS also yields consistent improvements in Dice scores over baselines on a cross-modality adult brain gliomas segmentation task composed of four different contrasts from the BraTS 2020 challenge dataset, where 95% of a target supervised model performance is reached when no target modality annotations are available. We report that 99% and 100% of this maximum performance can be attained if 20% and 50% of the target data is additionally annotated, which further demonstrates that MoDATTS can be leveraged to reduce the annotation burden.
引用
收藏
页数:16
相关论文
共 50 条
  • [31] Unsupervised Domain Adaptation for Remote Sensing Image Segmentation Based on Adversarial Learning and Self-Training
    Liang, Chenbin
    Cheng, Bo
    Xiao, Baihua
    Dong, Yunyun
    IEEE GEOSCIENCE AND REMOTE SENSING LETTERS, 2023, 20
  • [32] StAC-DA: Structure aware cross-modality domain adaptation framework with image and feature-level adaptation for medical image segmentation
    Baldeon-Calisto, Maria
    Lai-Yuen, Susana K.
    Puente-Mejia, Bernardo
    DIGITAL HEALTH, 2024, 10
  • [33] A lightweight image-level segmentation method for steel surface defects based on cross-layer feature fusion
    Wang, Peng
    Li, Liangliang
    Sha, Baolin
    Li, Xiaoyan
    Lue, Zhigang
    INSIGHT, 2024, 66 (03) : 167 - 173
  • [34] Trans-cGAN: transformer-Unet-based generative adversarial networks for cross-modality magnetic resonance image synthesis
    Yan Li
    Na Han
    Yuxiang Qin
    Jing Zhang
    Jinxia Su
    Statistics and Computing, 2023, 33
  • [35] LETCP: A Label-Efficient Transformer-Based Contrastive Pre-Training Method for Brain Tumor Segmentation
    Chen, Shoucun
    Zhang, Jing
    Zhang, Tianchi
    APPLIED SCIENCES-BASEL, 2022, 12 (21):
  • [36] Trans-cGAN: transformer-Unet-based generative adversarial networks for cross-modality magnetic resonance image synthesis
    Li, Yan
    Han, Na
    Qin, Yuxiang
    Zhang, Jing
    Su, Jinxia
    STATISTICS AND COMPUTING, 2023, 33 (05)
  • [37] GSAC-UFormer: Groupwise Self-Attention Convolutional Transformer-Based UNet for Medical Image Segmentation
    Garbaz, Anass
    Oukdach, Yassine
    Charfi, Said
    El Ansari, Mohamed
    Koutti, Lahcen
    Salihoun, Mouna
    COGNITIVE COMPUTATION, 2025, 17 (02)
  • [38] Label-efficient transformer-based framework with self-supervised strategies for heterogeneous lung tumor segmentation
    Liu, Zhenbing
    Li, Weixing
    Cui, Yanfen
    Chen, Xin
    Pan, Xipeng
    Ye, Guanchao
    Wu, Guangyao
    Liao, Yongde
    Volmer, Leroy
    Wee, Leonard
    Dekker, Andre
    Han, Chu
    Liu, Zaiyi
    Shi, Zhenwei
    EXPERT SYSTEMS WITH APPLICATIONS, 2025, 269
  • [39] Wavelet-based spectrum transfer with collaborative learning for unsupervised bidirectional cross-modality domain adaptation on medical image segmentation
    Liu, Shaolei
    Qu, Linhao
    Yin, Siqi
    Wang, Manning
    Song, Zhijian
    NEURAL COMPUTING & APPLICATIONS, 2024, 36 (12): : 6741 - 6755
  • [40] ST-IRGS: A Region-Based Self-Training Algorithm Applied to Hyperspectral Image Classification and Segmentation
    Li, Fan
    Clausi, David A.
    Xu, Linlin
    Wong, Alexander
    IEEE TRANSACTIONS ON GEOSCIENCE AND REMOTE SENSING, 2018, 56 (01): : 3 - 16