Automatic segmentation of echocardiographic images using a shifted windows vision transformer architecture

被引:0
|
作者
Nemri, Souha [1 ]
Duong, Luc [1 ]
机构
[1] Ecole Technol Super, Software & IT Engn Dept, Intervent Imaging Lab LIVE, 1100 Notre Dame St West, Montreal, PQ H3C 1K3, Canada
来源
基金
加拿大自然科学与工程研究理事会;
关键词
echocardiography; semantic segmentation; left ventricle; transformers; U-Net;
D O I
10.1088/2057-1976/ad7594
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Echocardiography is one the most commonly used imaging modalities for the diagnosis of congenital heart disease. Echocardiographic image analysis is crucial to obtaining accurate cardiac anatomy information. Semantic segmentation models can be used to precisely delimit the borders of the left ventricle, and allow an accurate and automatic identification of the region of interest, which can be extremely useful for cardiologists. In the field of computer vision, convolutional neural network (CNN) architectures remain dominant. Existing CNN approaches have proved highly efficient for the segmentation of various medical images over the past decade. However, these solutions usually struggle to capture long-range dependencies, especially when it comes to images with objects of different scales and complex structures. In this study, we present an efficient method for semantic segmentation of echocardiographic images that overcomes these challenges by leveraging the self-attention mechanism of the Transformer architecture. The proposed solution extracts long-range dependencies and efficiently processes objects at different scales, improving performance in a variety of tasks. We introduce Shifted Windows Transformer models (Swin Transformers), which encode both the content of anatomical structures and the relationship between them. Our solution combines the Swin Transformer and U-Net architectures, producing a U-shaped variant. The validation of the proposed method is performed with the EchoNet-Dynamic dataset used to train our model. The results show an accuracy of 0.97, a Dice coefficient of 0.87, and an Intersection over union (IoU) of 0.78. Swin Transformer models are promising for semantically segmenting echocardiographic images and may help assist cardiologists in automatically analyzing and measuring complex echocardiographic images.
引用
收藏
页数:10
相关论文
共 50 条
  • [21] Shift-ViT : Siamese Vision Transformer using Shifted Branches
    Aim, Dasom
    Kim, Hyeong Jin
    Kim, Sangwon
    Ko, Byoung Chul
    2022 37TH INTERNATIONAL TECHNICAL CONFERENCE ON CIRCUITS/SYSTEMS, COMPUTERS AND COMMUNICATIONS (ITC-CSCC 2022), 2022, : 259 - 261
  • [22] Medical images segmentation using gabor filters applied to echocardiographic images
    Bosnjak, A
    Montilla, A
    Torrealba, V
    COMPUTERS IN CARDIOLOGY 1998, VOL 25, 1998, 25 : 457 - 460
  • [23] Rethinking encoder-decoder architecture using vision transformer for colorectal polyp and surgical instruments segmentation
    Iqbal, Ahmed
    Ahmed, Zohair
    Usman, Muhammad
    Malik, Isra
    ENGINEERING APPLICATIONS OF ARTIFICIAL INTELLIGENCE, 2024, 136
  • [24] Swin-GAN: generative adversarial network based on shifted windows transformer architecture for image generation
    Shibin Wang
    Zidiao Gao
    Dong Liu
    The Visual Computer, 2023, 39 : 6085 - 6095
  • [25] Sediment grain segmentation in thin-section images using dual-modal Vision Transformer
    Zheng, Dongyu
    Hou, Li
    Hu, Xiumian
    Hou, Mingcai
    Dong, Kai
    Hu, Sihai
    Teng, Runlin
    Ma, Chao
    COMPUTERS & GEOSCIENCES, 2024, 191
  • [26] Swin-GAN: generative adversarial network based on shifted windows transformer architecture for image generation
    Wang, Shibin
    Gao, Zidiao
    Liu, Dong
    VISUAL COMPUTER, 2023, 39 (12): : 6085 - 6095
  • [27] An amalgamation of vision transformer with convolutional neural network for automatic lung tumor segmentation
    Tyagi, Shweta
    Kushnure, Devidas T.
    Talbar, Sanjay N.
    COMPUTERIZED MEDICAL IMAGING AND GRAPHICS, 2023, 108
  • [28] A vision transformer for emphysema classification using CT images
    Wu, Yanan
    Qi, Shouliang
    Sun, Yu
    Xia, Shuyue
    Yao, Yudong
    Qian, Wei
    PHYSICS IN MEDICINE AND BIOLOGY, 2021, 66 (24):
  • [29] Image Classification Using Vision Transformer for EtC Images
    Hamano, Genki
    Imaizumi, Shoko
    Kiya, Hitoshi
    PROCEEDINGS OF 2022 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2022, : 1506 - 1513
  • [30] Manipulation Detection in Satellite Images Using Vision Transformer
    Horvath, Janos
    Baireddy, Sriram
    Hao, Hanxiang
    Montserrat, Daniel Mas
    Delp, Edward J.
    2021 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION WORKSHOPS, CVPRW 2021, 2021, : 1032 - 1041