Image-level supervision and self-training for transformer-based cross-modality tumor segmentation

被引:0
|
作者
d'Assier, Malo Alefsen de Boisredon [1 ]
Portafaix, Aloys [1 ,2 ]
Vorontsov, Eugene [3 ]
Le, William Trung [1 ,2 ]
Kadoury, Samuel [1 ,2 ]
机构
[1] Polytech Montreal, Montreal, PQ, Canada
[2] Univ Montreal, Ctr Rech Ctr Hosp, Montreal, PQ, Canada
[3] Paige, Montreal, PQ, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Tumor Segmentation; Semi-supervised Learning; Domain adaptation; Self-training; UNSUPERVISED DOMAIN ADAPTATION;
D O I
10.1016/j.media.2024.103287
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks are commonly used for automated medical image segmentation, but models will frequently struggle to generalize well across different imaging modalities. This issue is particularly problematic due to the limited availability of annotated data, both in the target as well as the source modality, making it difficult to deploy these models on a larger scale. To overcome these challenges, we propose a new semi- supervised training strategy called MoDATTS. Our approach is designed for accurate cross-modality 3D tumor segmentation on unpaired bi-modal datasets. An image-to-image translation strategy between modalities is used to produce synthetic but annotated images and labels in the desired modality and improve generalization to the unannotated target modality. We also use powerful vision transformer architectures for both image translation (TransUNet) and segmentation (Medformer) tasks and introduce an iterative self-training procedure in the later task to further close the domain gap between modalities, thus also training on unlabeled images in the target modality. MoDATTS additionally allows the possibility to exploit image-level labels with a semi-supervised objective that encourages the model to disentangle tumors from the background. This semi-supervised methodology helps in particular to maintain downstream segmentation performance when pixel-level label scarcity is also present in the source modality dataset, or when the source dataset contains healthy controls. The proposed model achieves superior performance compared to other methods from participating teams in the CrossMoDA 2022 vestibular schwannoma (VS) segmentation challenge, as evidenced by its reported top Dice score of 0.87 +/- 0.04 . 87 +/- 0 . 04 for the VS segmentation. MoDATTS also yields consistent improvements in Dice scores over baselines on a cross-modality adult brain gliomas segmentation task composed of four different contrasts from the BraTS 2020 challenge dataset, where 95% of a target supervised model performance is reached when no target modality annotations are available. We report that 99% and 100% of this maximum performance can be attained if 20% and 50% of the target data is additionally annotated, which further demonstrates that MoDATTS can be leveraged to reduce the annotation burden.
引用
收藏
页数:16
相关论文
共 50 条
  • [21] PSIGAN: Joint Probabilistic Segmentation and Image Distribution Matching for Unpaired Cross-Modality Adaptation-Based MRI Segmentation
    Jiang, Jue
    Hu, Yu-Chi
    Tyagi, Neelam
    Rimner, Andreas
    Lee, Nancy
    Deasy, Joseph O.
    Berry, Sean
    Veeraraghavan, Harini
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2020, 39 (12) : 4071 - 4084
  • [22] TIG-UDA: Generative unsupervised domain adaptation with transformer-embedded invariance for cross-modality medical image segmentation
    Li, Jiapeng
    Chen, Yijia
    Li, Shijie
    Xu, Lisheng
    Qian, Wei
    Tian, Shuai
    Qi, Lin
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2025, 106
  • [23] On Efficient Transformer-Based Image Pre-training for Low-Level Vision
    Li, Wenbo
    Lu, Xin
    Qian, Shengju
    Lu, Jiangbo
    PROCEEDINGS OF THE THIRTY-SECOND INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2023, 2023, : 1089 - 1097
  • [24] Self-training and Multi-level Adversarial Network for Domain Adaptive Remote Sensing Image Segmentation
    Zheng, Yilin
    He, Lingmin
    Wu, Xiangping
    Pan, Chen
    NEURAL PROCESSING LETTERS, 2023, 55 (08) : 10613 - 10638
  • [25] Multi-Scale Transformer Network With Edge-Aware Pre-Training for Cross-Modality MR Image Synthesis
    Li, Yonghao
    Zhou, Tao
    He, Kelei
    Zhou, Yi
    Shen, Dinggang
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2023, 42 (11) : 3395 - 3407
  • [26] Self-training and Multi-level Adversarial Network for Domain Adaptive Remote Sensing Image Segmentation
    Yilin Zheng
    Lingmin He
    Xiangping Wu
    Chen Pan
    Neural Processing Letters, 2023, 55 : 10613 - 10638
  • [27] Self-training guided disentangled adaptation for cross-domain remote sensing image semantic segmentation
    Zhao, Qi
    Lyu, Shuchang
    Zhao, Hongbo
    Liu, Binghao
    Chen, Lijiang
    Cheng, Guangliang
    INTERNATIONAL JOURNAL OF APPLIED EARTH OBSERVATION AND GEOINFORMATION, 2024, 127
  • [28] Disentangled representation and cross-modality image translation based unsupervised domain adaptation method for abdominal organ segmentation
    Kaida Jiang
    Li Quan
    Tao Gong
    International Journal of Computer Assisted Radiology and Surgery, 2022, 17 : 1101 - 1113
  • [29] Disentangled representation and cross-modality image translation based unsupervised domain adaptation method for abdominal organ segmentation
    Jiang, Kaida
    Quan, Li
    Gong, Tao
    INTERNATIONAL JOURNAL OF COMPUTER ASSISTED RADIOLOGY AND SURGERY, 2022, 17 (06) : 1101 - 1113
  • [30] 10.1109/ICIPMC62364.2024.10586680 Unsupervised Domain Adaptation for Cross-Modality Cardiac Image Segmentation Based on Contrastive Image Synthesis
    Feng, Haoran
    Qu, Lei
    Wu, Jun
    Tian, Hao
    Zhang, Yong
    Li, Xiaohu
    2024 3RD INTERNATIONAL CONFERENCE ON IMAGE PROCESSING AND MEDIA COMPUTING, ICIPMC 2024, 2024, : 144 - 150