Performance Comparison of Vision Transformer-Based Models in Medical Image Classification

被引:1
|
作者
Kanca, Elif [1 ]
Ayas, Selen [2 ]
Kablan, Elif Baykal [1 ]
Ekinci, Murat [2 ]
机构
[1] Karadeniz Tech Univ, Yazilim Muhendisligi, Trabzon, Turkiye
[2] Karadeniz Tech Univ, Bilgisayar Muhendisligi, Trabzon, Turkiye
关键词
Vision transformer-based models; transformers; medical image classification;
D O I
10.1109/SIU59756.2023.10223892
中图分类号
TP18 [人工智能理论];
学科分类号
081104 ; 0812 ; 0835 ; 1405 ;
摘要
In recent years, convolutional neural networks have shown significant success and are frequently used in medical image analysis applications. However, the convolution process in convolutional neural networks limits learning of long-term pixel dependencies in the local receptive field. Inspired by the success of transformer architectures in encoding long-term dependencies and learning more efficient feature representation in natural language processing, publicly available color fundus retina, skin lesion, chest X-ray, and breast histology images are classified using Vision Transformer (ViT), Data-Efficient Transformer (DeiT), Swin Transformer, and Pyramid Vision Transformer v2 (PVTv2) models and their classification performances are compared in this study. The results show that the highest accuracy values are obtained with the DeiT model at 96.5% in the chest X-ray dataset, the PVTv2 model at 91.6% in the breast histology dataset, the PVTv2 model at 91.3% in the retina fundus dataset, and the Swin model at 91.0% in the skin lesion dataset.
引用
收藏
页数:4
相关论文
共 50 条
  • [31] A Vision Transformer-Based Approach to Bearing Fault Classification via Vibration Signals
    Zim, Abid Hasan
    Ashraf, Aeyan
    Iqbal, Aquib
    Malik, Asad
    Kuribayashi, Minoru
    PROCEEDINGS OF 2022 ASIA-PACIFIC SIGNAL AND INFORMATION PROCESSING ASSOCIATION ANNUAL SUMMIT AND CONFERENCE (APSIPA ASC), 2022, : 1321 - 1326
  • [32] A Survey of Transformer-Based Few-Shot Image Classification Techniques
    Song, Chaoqi
    Liu, Ying
    He, Jinglu
    2024 6TH INTERNATIONAL CONFERENCE ON NATURAL LANGUAGE PROCESSING, ICNLP 2024, 2024, : 599 - 608
  • [33] Efficient Image Captioning Based on Vision Transformer Models
    Elbedwehy, Samar
    Medhat, T.
    Hamza, Taher
    Alrahmawy, Mohammed F.
    CMC-COMPUTERS MATERIALS & CONTINUA, 2022, 73 (01): : 1483 - 1500
  • [34] Medical image classification: Knowledge transfer via residual U-Net and vision transformer-based teacher-student model with knowledge distillation
    Song, Yucheng
    Wang, Jincan
    Ge, Yifan
    Li, Lifeng
    Guo, Jia
    Dong, Quanxing
    Liao, Zhifang
    JOURNAL OF VISUAL COMMUNICATION AND IMAGE REPRESENTATION, 2024, 102
  • [35] Transformer-Based Annotation Bias-Aware Medical Image Segmentation
    Liao, Zehui
    Hu, Shishuai
    Xie, Yutong
    Xia, Yong
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT IV, 2023, 14223 : 24 - 34
  • [36] REVIEW OF VISION TRANSFORMER MODELS FOR REMOTE SENSING IMAGE SCENE CLASSIFICATION
    Lv, Pengyuan
    Wu, Wenjun
    Zhong, Yanfei
    Zhang, Liangpei
    2022 IEEE INTERNATIONAL GEOSCIENCE AND REMOTE SENSING SYMPOSIUM (IGARSS 2022), 2022, : 2231 - 2234
  • [37] PtychoDV: Vision Transformer-Based Deep Unrolling Network for Ptychographic Image Reconstruction
    Gan, Weijie
    Zhai, Qiuchen
    Mccann, Michael T.
    Cardona, Cristina Garcia
    Kamilov, Ulugbek S.
    Wohlberg, Brendt
    IEEE OPEN JOURNAL OF SIGNAL PROCESSING, 2024, 5 : 539 - 547
  • [38] Tweets Topic Classification and Sentiment Analysis Based on Transformer-Based Language Models
    Mandal, Ranju
    Chen, Jinyan
    Becken, Susanne
    Stantic, Bela
    VIETNAM JOURNAL OF COMPUTER SCIENCE, 2023, 10 (02) : 117 - 134
  • [39] TanrsColour: Transformer-based medical image colourization with content and structure preservation
    Liu, Qinghai
    Zhao, Dengping
    Tang, Lun
    Xu, Limin
    IET IMAGE PROCESSING, 2024, 18 (10) : 2702 - 2714
  • [40] An Unsupervised Method for Industrial Image Anomaly Detection with Vision Transformer-Based Autoencoder
    Yang, Qiying
    Guo, Rongzuo
    SENSORS, 2024, 24 (08)