Enhance fashion classification of mosquito vector species via self-supervised vision transformer

被引:0
|
作者
Kittichai, Veerayuth [1 ]
Kaewthamasorn, Morakot [2 ]
Chaiphongpachara, Tanawat [3 ]
Laojun, Sedthapong [3 ]
Saiwichai, Tawee [4 ]
Naing, Kaung Myat [6 ]
Tongloy, Teerawat [6 ]
Boonsang, Siridech [5 ]
Chuwongin, Santhad [6 ]
机构
[1] King Mongkuts Inst Technol Ladkrabang, Fac Med, Bangkok, Thailand
[2] Chulalongkorn Univ, Fac Vet Sci, Vet Parasitol Res Unit, Bangkok, Thailand
[3] Suan Sunandha Rajabhat Univ, Coll Allied Hlth Sci, Dept Publ Hlth & Hlth Promot, Bangkok, Thailand
[4] Mahidol Univ, Fac Publ Hlth, Dept Parasitol & Entomol, Nakhon Pathom, Thailand
[5] King Mongkuts Inst Technol Ladkrabang, Sch Engn, Dept Elect Engn, Bangkok, Thailand
[6] King Mongkuts Inst Technol Ladkrabang, Coll Adv Mfg Innovat, Bangkok, Thailand
来源
SCIENTIFIC REPORTS | 2024年 / 14卷 / 01期
关键词
Mosquito vector species; Artificial intelligence; Self-distillation with unlabeled data; Mobile phone application;
D O I
10.1038/s41598-024-83358-8
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Vector-borne diseases pose a major worldwide health concern, impacting more than 1 billion people globally. Among various blood-feeding arthropods, mosquitoes stand out as the primary carriers of diseases significant in both medical and veterinary fields. Hence, comprehending their distinct role fulfilled by different mosquito types is crucial for efficiently addressing and enhancing control measures against mosquito-transmitted diseases. The conventional method for identifying mosquito species is laborious and requires significant effort to learn. Classification is subsequently carried out by skilled laboratory personnel, rendering the process inherently time-intensive and restricting the task to entomology specialists. Therefore, integrating artificial intelligence with standard taxonomy, such as molecular techniques, is essential for accurate mosquito species identification. Advancement in novel tools with artificial intelligence has challenged the task of developing an automated system for sample collection and identification. This study aims to introduce a self-supervised Vision Transformer supporting an automatic model for classifying mosquitoes found across various regions of Thailand. The objective is to utilize self-distillation with unlabeled data (DINOv2) to develop models on a mobile phone-captured dataset containing 16 species of female mosquitoes, including those known for transmitting malaria and dengue. The DINOv2 model surpassed the ViT baseline model in precision and recall for all mosquito species. When compared on a species-specific level, utilizing the DINOv2 model resulted in reductions in false negatives and false positives, along with enhancements in precision and recall values, in contrast to the baseline model, across all mosquito species. Notably, at least 10 classes exhibited outstanding performance, achieving above precision and recall rates exceeding 90%. Remarkably, when applying cropping techniques to the dataset instead of utilizing the original photographs, there was a significant improvement in performance across all DINOv2 models studied. This is demonstrated by an increase in recall to 87.86%, precision to 91.71%, F1 score to 88.71%, and accuracy to 98.45%, respectively. Malaria mosquito species can be easily distinguished from another genus like Aedes, Mansonia, Armigeres, and Culex, respectively. While classifying malaria vector species presented challenges for the DINOv2 model, utilizing the cropped images enhanced precision by up to 96% for identifying one of the top three malaria vectors in Thailand, Anopheles minimus. A proficiently trained DINOv2 model, coupled with effective data management, can contribute to the development of a mobile phone application. Furthermore, this method shows promise in supporting field professionals who are not entomology experts in effectively addressing pathogens responsible for diseases transmitted by female mosquitoes.
引用
收藏
页数:16
相关论文
共 50 条
  • [1] Positional Label for Self-Supervised Vision Transformer
    Zhang, Zhemin
    Gong, Xun
    THIRTY-SEVENTH AAAI CONFERENCE ON ARTIFICIAL INTELLIGENCE, VOL 37 NO 3, 2023, : 3516 - 3524
  • [2] Histopathological Image Classification based on Self-Supervised Vision Transformer and Weak Labels
    Gul, Ahmet Gokberk
    Cetin, Oezdemir
    Reich, Christoph
    Flinner, Nadine
    Prangemeier, Tim
    Koeppl, Heinz
    MEDICAL IMAGING 2022: DIGITAL AND COMPUTATIONAL PATHOLOGY, 2022, 12039
  • [3] Enhancing mosquito classification through self-supervised learning
    Charoenpanyakul, Ratana
    Kittichai, Veerayuth
    Eiamsamang, Songpol
    Sriwichai, Patchara
    Pinetsuksai, Natchapon
    Naing, Kaung Myat
    Tongloy, Teerawat
    Boonsang, Siridech
    Chuwongin, Santhad
    SCIENTIFIC REPORTS, 2024, 14 (01):
  • [4] MonoViT: Self-Supervised Monocular Depth Estimation with a Vision Transformer
    Zhao, Chaoqiang
    Zhang, Youmin
    Poggi, Matteo
    Tosi, Fabio
    Guo, Xianda
    Zhu, Zheng
    Huang, Guan
    Tang, Yang
    Mattoccia, Stefano
    2022 INTERNATIONAL CONFERENCE ON 3D VISION, 3DV, 2022, : 668 - 678
  • [5] Multi-scale vision transformer classification model with self-supervised learning and dilated convolution
    Xing, Liping
    Jin, Hongmei
    Li, Hong-an
    Li, Zhanli
    COMPUTERS & ELECTRICAL ENGINEERING, 2022, 103
  • [6] Multimodal Image Fusion via Self-Supervised Transformer
    Zhang, Jing
    Liu, Yu
    Liu, Aiping
    Xie, Qingguo
    Ward, Rabab
    Wang, Z. Jane
    Chen, Xun
    IEEE SENSORS JOURNAL, 2023, 23 (09) : 9796 - 9807
  • [7] Self-Supervised Transformer Networks for Error Classification of Tightening Traces
    Wilkman, Dennis Bogatov
    Tang, Lifei
    Morozovska, Kateryna
    Bragone, Federica
    2022 21ST IEEE INTERNATIONAL CONFERENCE ON MACHINE LEARNING AND APPLICATIONS, ICMLA, 2022, : 1373 - 1380
  • [8] Self-supervised Video Transformer
    Ranasinghe, Kanchana
    Naseer, Muzammal
    Khan, Salman
    Khan, Fahad Shahbaz
    Ryoo, Michael S.
    2022 IEEE/CVF CONFERENCE ON COMPUTER VISION AND PATTERN RECOGNITION (CVPR 2022), 2022, : 2864 - 2874
  • [9] IMPROVING ACOUSTIC SCENE CLASSIFICATION VIA SELF-SUPERVISED AND SEMI-SUPERVISED LEARNING WITH EFFICIENT AUDIO TRANSFORMER
    Liang, Yuzhe
    Chen, Wenxi
    Jiang, Anbai
    Qiu, Yihong
    Zhen, Xinhu
    Huang, Wen
    Han, Bing
    Qian, Yanmin
    Fang, Pingyi
    Zhang, Wei-Qiang
    Lu, Cheng
    Liu, Jia
    Chen, Xie
    2024 IEEE INTERNATIONAL CONFERENCE ON MULTIMEDIA AND EXPO WORKSHOPS, ICMEW 2024, 2024,
  • [10] Self-supervised Vision Transformer are Scalable Generative Models for Domain Generalization
    Doerrich, Sebastian
    Di Salvo, Francesco
    Ledig, Christian
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION - MICCAI 2024, PT X, 2024, 15010 : 644 - 654