Image-level supervision and self-training for transformer-based cross-modality tumor segmentation

被引:0
|
作者
d'Assier, Malo Alefsen de Boisredon [1 ]
Portafaix, Aloys [1 ,2 ]
Vorontsov, Eugene [3 ]
Le, William Trung [1 ,2 ]
Kadoury, Samuel [1 ,2 ]
机构
[1] Polytech Montreal, Montreal, PQ, Canada
[2] Univ Montreal, Ctr Rech Ctr Hosp, Montreal, PQ, Canada
[3] Paige, Montreal, PQ, Canada
基金
加拿大自然科学与工程研究理事会;
关键词
Tumor Segmentation; Semi-supervised Learning; Domain adaptation; Self-training; UNSUPERVISED DOMAIN ADAPTATION;
D O I
10.1016/j.media.2024.103287
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Deep neural networks are commonly used for automated medical image segmentation, but models will frequently struggle to generalize well across different imaging modalities. This issue is particularly problematic due to the limited availability of annotated data, both in the target as well as the source modality, making it difficult to deploy these models on a larger scale. To overcome these challenges, we propose a new semi- supervised training strategy called MoDATTS. Our approach is designed for accurate cross-modality 3D tumor segmentation on unpaired bi-modal datasets. An image-to-image translation strategy between modalities is used to produce synthetic but annotated images and labels in the desired modality and improve generalization to the unannotated target modality. We also use powerful vision transformer architectures for both image translation (TransUNet) and segmentation (Medformer) tasks and introduce an iterative self-training procedure in the later task to further close the domain gap between modalities, thus also training on unlabeled images in the target modality. MoDATTS additionally allows the possibility to exploit image-level labels with a semi-supervised objective that encourages the model to disentangle tumors from the background. This semi-supervised methodology helps in particular to maintain downstream segmentation performance when pixel-level label scarcity is also present in the source modality dataset, or when the source dataset contains healthy controls. The proposed model achieves superior performance compared to other methods from participating teams in the CrossMoDA 2022 vestibular schwannoma (VS) segmentation challenge, as evidenced by its reported top Dice score of 0.87 +/- 0.04 . 87 +/- 0 . 04 for the VS segmentation. MoDATTS also yields consistent improvements in Dice scores over baselines on a cross-modality adult brain gliomas segmentation task composed of four different contrasts from the BraTS 2020 challenge dataset, where 95% of a target supervised model performance is reached when no target modality annotations are available. We report that 99% and 100% of this maximum performance can be attained if 20% and 50% of the target data is additionally annotated, which further demonstrates that MoDATTS can be leveraged to reduce the annotation burden.
引用
收藏
页数:16
相关论文
共 50 条
  • [41] Masked Image Modeling Meets Self-Distillation: A Transformer-Based Prostate Gland Segmentation Framework for Pathology Slides
    Zhang, Haoyue
    Patkar, Sushant
    Lis, Rosina
    Merino, Maria J.
    Pinto, Peter A.
    Choyke, Peter L.
    Turkbey, Baris
    Harmon, Stephanie
    CANCERS, 2024, 16 (23)
  • [42] FPL plus : Filtered Pseudo Label-Based Unsupervised Cross-Modality Adaptation for 3D Medical Image Segmentation
    Wu, Jianghao
    Guo, Dong
    Wang, Guotai
    Yue, Qiang
    Yu, Huijun
    Li, Kang
    Zhang, Shaoting
    IEEE TRANSACTIONS ON MEDICAL IMAGING, 2024, 43 (09) : 3098 - 3109
  • [43] RETRACTED: Artificial Intelligence-Based Semisupervised Self-Training Algorithm in Pathological Tissue Image Segmentation (Retracted Article)
    Li, Qun
    Liu, Linlin
    COMPUTATIONAL INTELLIGENCE AND NEUROSCIENCE, 2022, 2022
  • [44] Local contrastive loss with pseudo-label based self-training for semi-supervised medical image segmentation
    Chaitanya, Krishna
    Erdil, Ertunc
    Karani, Neerav
    Konukoglu, Ender
    MEDICAL IMAGE ANALYSIS, 2023, 87
  • [45] SAM-guided contrast based self-training for source-free cross-domain semantic segmentation
    Ren, Qinghua
    Hou, Ke
    Zhan, Yongzhao
    Wang, Chen
    MULTIMEDIA SYSTEMS, 2024, 30 (04)
  • [46] Intra-modality masked image modeling: A self-supervised pre-training method for brain tumor segmentation
    Qi, Liangce
    Shi, Weili
    Miao, Yu
    Li, Yonghui
    Feng, Guanyuan
    Jiang, Zhengang
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2024, 95
  • [47] Self-training Based Semi-Supervised Learning and U-Net with Denoiser for Teeth Segmentation in X-Ray Image
    Lin, Zhouhao
    Yang, Yibo
    Huang, Anrui
    Shou, Zeyang
    Zhang, Qizhong
    SEMI-SUPERVISED TOOTH SEGMENTATION, SEMITOOTHSEG 2023, 2025, 14623 : 124 - 132
  • [48] A dual-branch selection method with pseudo-label based self-training for semi-supervised smoke image segmentation
    Li, Haibin
    Qi, Jiawei
    Li, Yaqian
    Zhang, Wenming
    DIGITAL SIGNAL PROCESSING, 2024, 145
  • [49] BTSwin-Unet: 3D U-shaped Symmetrical Swin Transformer-based Network for Brain Tumor Segmentation with Self-supervised Pre-training
    Junjie Liang
    Cihui Yang
    Jingting Zhong
    Xiaoli Ye
    Neural Processing Letters, 2023, 55 : 3695 - 3713
  • [50] BTSwin-Unet: 3D U-shaped Symmetrical Swin Transformer-based Network for Brain Tumor Segmentation with Self-supervised Pre-training
    Liang, Junjie
    Yang, Cihui
    Zhong, Jingting
    Ye, Xiaoli
    NEURAL PROCESSING LETTERS, 2023, 55 (04) : 3695 - 3713