PEMMA: Parameter-Efficient Multi-Modal Adaptation for Medical Image Segmentation

被引:0
|
作者
Saadi, Nada [1 ]
Saeed, Numan [1 ]
Yaqub, Mohammad [1 ]
Nandakumar, Karthik [1 ]
机构
[1] Mohamed Bin Zayed Univ Artificial Intelligence, Abu Dhabi, U Arab Emirates
关键词
Multi-modal Adaptation; Low-rank Adaptation; Parameter-Efficiency; Cross-modal Entanglement; 3D Medical Image Segmentation;
D O I
10.1007/978-3-031-72390-2_25
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
Imaging modalities such as Computed Tomography (CT) and Positron Emission Tomography (PET) are key in cancer detection, inspiring Deep Neural Networks (DNN) models that merge these scans for tumor segmentation. When both CT and PET scans are available, it is common to combine them as two channels of the input to the segmentation model. However, this method requires both scan types during training and inference, posing a challenge due to the limited availability of PET scans, thereby sometimes limiting the process to CT scans only. Hence, there is a need to develop a flexible DNN architecture that can be trained/updated using only CT scans but can effectively utilize PET scans when they become available. In this work, we propose a parameter-efficient multi-modal adaptation (PEMMA) framework for lightweight upgrading of a transformer-based segmentation model trained only on CT scans to also incorporate PET scans. The benefits of the proposed approach are two-fold. Firstly, we leverage the inherent modularity of the transformer architecture and perform low-rank adaptation (LoRA) of the attention weights to achieve parameter-efficient adaptation. Secondly, since the PEMMA framework attempts to minimize cross-modal entanglement, it is possible to subsequently update the combined model using only one modality, without causing catastrophic forgetting of the other modality. Our proposed method achieves comparable results with the performance of early fusion techniques with just 8% of the trainable parameters, especially with a remarkable +28% improvement on the average dice score on PET scans when trained on a single modality.
引用
收藏
页码:262 / 271
页数:10
相关论文
共 50 条
  • [31] Adherent Peanut Image Segmentation Based on Multi-Modal Fusion
    Wang, Yujing
    Ye, Fang
    Zeng, Jiusun
    Cai, Jinhui
    Huang, Wangsen
    SENSORS, 2024, 24 (14)
  • [32] Toward Unpaired Multi-modal Medical Image Segmentation via Learning Structured Semantic Consistency
    Yang, Jie
    Zhu, Ye
    Wang, Chaoqun
    Li, Zhen
    Zhang, Ruimao
    MEDICAL IMAGING WITH DEEP LEARNING, VOL 227, 2023, 227 : 1602 - 1622
  • [33] Parameter-Efficient Tuning with Special Token Adaptation
    Yang, Xiaocong
    Huang, James Y.
    Zhou, Wenxuan
    Chen, Muhao
    17TH CONFERENCE OF THE EUROPEAN CHAPTER OF THE ASSOCIATION FOR COMPUTATIONAL LINGUISTICS, EACL 2023, 2023, : 865 - 872
  • [34] Multi-modal medical image segmentation based on vector-valued active contour models
    Fang, Lingling
    Wang, Xin
    Wang, Lujie
    INFORMATION SCIENCES, 2020, 513 : 504 - 518
  • [35] Overview of Multi-Modal Brain Tumor MR Image Segmentation
    Zhang, Wenyin
    Wu, Yong
    Yang, Bo
    Hu, Shunbo
    Wu, Liang
    Dhelim, Sahraoui
    HEALTHCARE, 2021, 9 (08)
  • [36] Multi-modal brain tumor image segmentation based on SDAE
    Ding, Yi
    Dong, Rongfeng
    Lan, Tian
    Li, Xuerui
    Shen, Guangyu
    Chen, Hao
    Qin, Zhiguang
    INTERNATIONAL JOURNAL OF IMAGING SYSTEMS AND TECHNOLOGY, 2018, 28 (01) : 38 - 47
  • [37] Parameter-Efficient Model Adaptation for Vision Transformers
    He, Xuehai
    Li, Chuanyuan
    Zhang, Pengchuan
    Yang, Jianwei
    Wang, Xin Eric
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 1, 2023, : 817 - 825
  • [38] Dynamic Brightness Adaptation for Robust Multi-modal Image Fusion
    Sun, Yiming
    Cao, Bing
    Zhu, Pengfei
    Hu, Qinghua
    PROCEEDINGS OF THE THIRTY-THIRD INTERNATIONAL JOINT CONFERENCE ON ARTIFICIAL INTELLIGENCE, IJCAI 2024, 2024, : 1317 - 1325
  • [39] Efficient Parameter-Free Adaptive Multi-Modal Hashing
    Zheng, Chaoqun
    Zhu, Lei
    Zhang, Shusen
    Zhang, Huaxiang
    IEEE SIGNAL PROCESSING LETTERS, 2020, 27 (27) : 1270 - 1274
  • [40] Polyhistor: Parameter-Efficient Multi-Task Adaptation for Dense Vision Tasks
    Liu, Yen-Cheng
    Ma, Chih-Yao
    Tian, Junjiao
    He, Zijian
    Kira, Zsolt
    ADVANCES IN NEURAL INFORMATION PROCESSING SYSTEMS 35 (NEURIPS 2022), 2022,