Development and Validation of a Deep-Learning Network for Detecting Congenital Heart Disease from Multi-View Multi-Modal Transthoracic Echocardiograms

被引:3
|
作者
Cheng, Mingmei [1 ]
Wang, Jing [2 ,3 ]
Liu, Xiaofeng [4 ,5 ]
Wang, Yanzhong [6 ]
Wu, Qun [2 ]
Wang, Fangyun [2 ]
Li, Pei
Wang, Binbin [7 ,8 ,9 ]
Zhang, Xin [2 ]
Xie, Wanqing [1 ]
机构
[1] Anhui Med Univ, Sch Biomed Engn, Sch Mental Hlth & Psychol Sci, Dept Intelligent Med Engn, Hefei 230011, Peoples R China
[2] Capital Med Univ, Natl Ctr Childrens Hlth, Beijing Childrens Hosp, Heart Ctr, Beijing, Peoples R China
[3] Capital Med Univ, Sch Basic Med Sci, Beijing, Peoples R China
[4] Harvard Med Sch, Gordon Ctr Med Imaging, Boston, MA 02114 USA
[5] Massachusetts Gen Hosp, Boston, MA 02114 USA
[6] Kings Coll London, Fac Life Sci & Med, Sch Life Course & Populat Sci, London, England
[7] Natl Res Inst Family Planning, Ctr Genet, Beijing 100730, Peoples R China
[8] Peking Union Med Coll, Grad Sch, Beijing 100730, Peoples R China
[9] Harvard Univ, Beth Israel Deaconess Med Ctr, Harvard Med Sch, Boston, MA 02215 USA
基金
中国国家自然科学基金; 北京市自然科学基金;
关键词
SURGERY;
D O I
10.34133/research.0319
中图分类号
O [数理科学和化学]; P [天文学、地球科学]; Q [生物科学]; N [自然科学总论];
学科分类号
07 ; 0710 ; 09 ;
摘要
Early detection and treatment of congenital heart disease (CHD) can significantly improve the prognosis of children. However, inexperienced sonographers often face difficulties in recognizing CHD through transthoracic echocardiogram (TTE) images. In this study, 2 -dimensional (2D) and Doppler TTEs of children collected from 2 clinical groups from Beijing Children's Hospital between 2018 and 2022 were analyzed, including views of apical 4 chamber, subxiphoid long-axis view of 2 atria, parasternal long-axis view of the left ventricle, parasternal short-axis view of aorta, and suprasternal long-axis view. A deep learning (DL) framework was developed to identify cardiac views, integrate information from various views and modalities, visualize the high-risk region, and predict the probability of the subject being normal or having an atrial septal defect (ASD) or a ventricular septaldefect (VSD). A total of 1,932 children (1,255 healthy controls, 292 ASDs, and 385 VSDs) were collected from 2 clinical groups. For view classification, the DL model reached a mean [SD] accuracy of 0.989 [0.001]. For CHD screening, the model using both 2D and Doppler TTEs with 5 views achieved a mean [SD] area under the receiver operating characteristic curve (AUC) of 0.996 [0.000] and an accuracy of 0.994 [0.002] for within -center evaluation while reaching a mean [SD] AUC of 0.990 [0.003] and an accuracy of 0.993 [0.001] for cross-center test set. For the classification of healthy, ASD, and VSD, the model reached the mean [SD] accuracy of 0.991 [0.002] and 0.986 [0.001] for within- and cross-center evaluation, respectively. The DL models aggregating TTEs with more modalities and scanning views attained superior performance to approximate that of experienced sonographers. The incorporation of multiple views and modalities of TTEs in the model enables accurate identification of children with CHD in a noninvasive manner, suggesting the potential to enhance CHD detection performance and simplify the screening process.
引用
收藏
页数:9
相关论文
共 50 条
  • [1] Automated interpretation of congenital heart disease from multi-view echocardiograms
    Wang, Jing
    Liu, Xiaofeng
    Wang, Fangyun
    Zheng, Lin
    Gao, Fengqiao
    Zhang, Hanwen
    Zhang, Xin
    Xie, Wanqing
    Wang, Binbin
    MEDICAL IMAGE ANALYSIS, 2021, 69
  • [2] Deep Multi-View Representation Learning for Multi-modal Features of the Schizophrenia and Schizo-affective Disorder
    Qi, Jun
    Tejedor, Javier
    2016 IEEE INTERNATIONAL CONFERENCE ON ACOUSTICS, SPEECH AND SIGNAL PROCESSING PROCEEDINGS, 2016, : 952 - 956
  • [3] Patch-based deep multi-modal learning framework for Alzheimer's disease diagnosis using multi-view neuroimaging
    Liu, Fangyu
    Yuan, Shizhong
    Li, Weimin
    Xu, Qun
    Sheng, Bin
    BIOMEDICAL SIGNAL PROCESSING AND CONTROL, 2023, 80
  • [4] Detecting glaucoma from multi-modal data using probabilistic deep learning
    Huang, Xiaoqin
    Sun, Jian
    Gupta, Krati
    Montesano, Giovanni
    Crabb, David P.
    Garway-Heath, David F.
    Brusini, Paolo
    Lanzetta, Paolo
    Oddone, Francesco
    Turpin, Andrew
    McKendrick, Allison M.
    Johnson, Chris A.
    Yousefi, Siamak
    FRONTIERS IN MEDICINE, 2022, 9
  • [5] Deep unsupervised multi-modal fusion network for detecting driver distraction
    Zhang, Yuxin
    Chen, Yiqiang
    Gao, Chenlong
    NEUROCOMPUTING, 2021, 421 : 26 - 38
  • [6] Deep unsupervised multi-modal fusion network for detecting driver distraction
    Zhang Y.
    Chen Y.
    Gao C.
    Neurocomputing, 2021, 421 : 26 - 38
  • [7] Multi-view Multi-modal Person Authentication from a Single Walking Image Sequence
    Muramatsu, Daigo
    Iwama, Haruyuki
    Makihara, Yasushi
    Yagi, Yasushi
    2013 INTERNATIONAL CONFERENCE ON BIOMETRICS (ICB), 2013,
  • [8] Deep cross-view autoencoder network for multi-view learning
    Mi, Jian-Xun
    Fu, Chang-Qing
    Chen, Tao
    Gou, Tingting
    MULTIMEDIA TOOLS AND APPLICATIONS, 2022, 81 (17) : 24645 - 24664
  • [9] Deep cross-view autoencoder network for multi-view learning
    Jian-Xun Mi
    Chang-Qing Fu
    Tao Chen
    Tingting Gou
    Multimedia Tools and Applications, 2022, 81 : 24645 - 24664
  • [10] Collaborative recommendation model based on multi-modal multi-view attention network: Movie and literature cases
    Hu, Zheng
    Cai, Shi-Min
    Wang, Jun
    Zhou, Tao
    APPLIED SOFT COMPUTING, 2023, 144