Benchmarking Deep Learning Models for Tooth Structure Segmentation

被引:14
|
作者
Schneider, L. [1 ,2 ]
Arsiwala-Scheppach, L. [1 ,2 ]
Krois, J. [1 ,2 ]
Meyer-Lueckel, H. [3 ]
Bressem, K. K. [4 ,5 ]
Niehues, S. M. [4 ]
Schwendicke, F. [1 ,2 ]
机构
[1] Charite, Dept Oral Diagnost Digital Hlth & Hlth Serv Res, Assmannshauser Str 4 6, D-14197 Berlin, Germany
[2] ITU WHO Focus Grp Hlth, Top Grp Dent Diagnost & Digital Dent, Geneva, Switzerland
[3] Univ Bern, Zahnmed Kliniken Univ Bern, Dept Restorat Prevent & Pediat Dent, Bern, Switzerland
[4] Charite Univ Med Berlin, Radiol Klin, Berlin, Germany
[5] Charite Univ Med Berlin, Berlin Inst Hlth, Berlin, Germany
关键词
computer vision; artificial intelligence; segmentation; tooth structures; transfer learning; neural networks;
D O I
10.1177/00220345221100169
中图分类号
R78 [口腔科学];
学科分类号
1003 ;
摘要
A wide range of deep learning (DL) architectures with varying depths are available, with developers usually choosing one or a few of them for their specific task in a nonsystematic way. Benchmarking (i.e., the systematic comparison of state-of-the art architectures on a specific task) may provide guidance in the model development process and may allow developers to make better decisions. However, comprehensive benchmarking has not been performed in dentistry yet. We aimed to benchmark a range of architecture designs for 1 specific, exemplary case: tooth structure segmentation on dental bitewing radiographs. We built 72 models for tooth structure (enamel, dentin, pulp, fillings, crowns) segmentation by combining 6 different DL network architectures (U-Net, U-Net++, Feature Pyramid Networks, LinkNet, Pyramid Scene Parsing Network, Mask Attention Network) with 12 encoders from 3 different encoder families (ResNet, VGG, DenseNet) of varying depth (e.g., VGG13, VGG16, VGG19). On each model design, 3 initialization strategies (ImageNet, CheXpert, random initialization) were applied, resulting overall into 216 trained models, which were trained up to 200 epochs with the Adam optimizer (learning rate = 0.0001) and a batch size of 32. Our data set consisted of 1,625 human-annotated dental bitewing radiographs. We used a 5-fold cross-validation scheme and quantified model performances primarily by the F1-score. Initialization with ImageNet or CheXpert weights significantly outperformed random initialization (P < 0.05). Deeper and more complex models did not necessarily perform better than less complex alternatives. VGG-based models were more robust across model configurations, while more complex models (e.g., from the ResNet family) achieved peak performances. In conclusion, initializing models with pretrained weights may be recommended when training models for dental radiographic analysis. Less complex model architectures may be competitive alternatives if computational resources and training time are restricting factors. Models developed and found superior on nondental data sets may not show this behavior for dental domain-specific tasks.
引用
收藏
页码:1343 / 1349
页数:7
相关论文
共 50 条
  • [11] Benchmarking Deep Learning Models for Automatic Ultrasonic Imaging Inspection
    Ye, Jiaxing
    Toyama, Nobuyuki
    IEEE ACCESS, 2021, 9 (09): : 36986 - 36994
  • [12] Architecture Analysis and Benchmarking of 3D U-Shaped Deep Learning Models for Thoracic Anatomical Segmentation
    Harirpoush, Arash
    Rasoulian, Amirhossein
    Kersten-Oertel, Marta
    Xiao, Yiming
    IEEE ACCESS, 2024, 12 : 127592 - 127603
  • [13] ROOD-MRI: Benchmarking the robustness of deep learning segmentation models to out-of-distribution and corrupted data in MRI
    Boone, Lyndon
    Biparva, Mahdi
    Forooshani, Parisa Mojiri
    Ramirez, Joel
    Masellis, Mario
    Bartha, Robert
    Symons, Sean
    Strother, Stephen
    Black, Sandra E.
    Heyn, Chris
    Martel, Anne L.
    Swartz, Richard H.
    Goubran, Maged
    NEUROIMAGE, 2023, 278
  • [14] Automatic Image Annotation and Deep Learning for Tooth CT Image Segmentation
    Gou, Miao
    Rao, Yunbo
    Zhang, Minglu
    Sun, Jianxun
    Cheng, Keyang
    IMAGE AND GRAPHICS, ICIG 2019, PT II, 2019, 11902 : 519 - 528
  • [15] EM-stellar: benchmarking deep learning for electron microscopy image segmentation
    Khadangi, Afshin
    Boudier, Thomas
    Rajagopal, Vijay
    BIOINFORMATICS, 2021, 37 (01) : 97 - 106
  • [16] Comparative studies of deep learning segmentation models for left ventricle segmentation
    Shoaib, Muhammad Ali
    Lai, Khin Wee
    Chuah, Joon Huang
    Hum, Yan Chai
    Ali, Raza
    Dhanalakshmi, Samiappan
    Wang, Huanhuan
    Wu, Xiang
    FRONTIERS IN PUBLIC HEALTH, 2022, 10
  • [17] Benchmarking the Robustness of Instance Segmentation Models
    Dalva, Yusuf
    Pehlivan, Hamza
    Altindis, Said Fahri
    Dundar, Aysegul
    IEEE TRANSACTIONS ON NEURAL NETWORKS AND LEARNING SYSTEMS, 2024, 35 (12) : 17021 - 17035
  • [18] Semantic Segmentation of Building Models with Deep Learning in CityGML
    Rashidan, Hanis
    Musliman, Ivin Amri
    Rahman, Alias Abdul
    Coors, Volker
    Buyuksalih, Gurcan
    19TH 3D GEOINFO CONFERENCE 2024, VOL. 48-4, 2024, : 97 - 102
  • [19] Certification of Deep Learning Models for Medical Image Segmentation
    Laousy, Othmane
    Araujo, Alexandre
    Chassagnon, Guillaume
    Paragios, Nikos
    Revel, Marie-Pierre
    Vakalopoulou, Maria
    MEDICAL IMAGE COMPUTING AND COMPUTER ASSISTED INTERVENTION, MICCAI 2023, PT IV, 2023, 14223 : 611 - 621
  • [20] Influence of imperfect annotations on deep learning segmentation models
    Brueckner, Christopher
    Liu, Chang
    Rist, Leonhard
    Maier, Andreas
    BILDVERARBEITUNG FUR DIE MEDIZIN 2024, 2024, : 226 - 231