Artificial intelligence in global health equity: an evaluation and discussion on the application of ChatGPT, in the Chinese National Medical Licensing Examination

被引:6
|
作者
Tong, Wenting [1 ]
Guan, Yongfu [2 ]
Chen, Jinping [2 ]
Huang, Xixuan [3 ]
Zhong, Yuting [4 ]
Zhang, Changrong [5 ]
Zhang, Hui [2 ,6 ]
机构
[1] Gannan Healthcare Vocat Coll, Dept Pharm, Ganzhou, Peoples R China
[2] Gannan Hlth Vocat Coll, Dept Rehabil & Elderly Care, Ganzhou, Jiangxi, Peoples R China
[3] Xiamen Univ, Dept Math, Xiamen, Fujian, Peoples R China
[4] Gannan Med Univ, Dept Anesthesiol, Ganzhou, Jiangxi, Peoples R China
[5] Qinghai Univ, Dept Chinese Med, Affiliated Hosp, Xining, Qinghai, Peoples R China
[6] Univ Roma Tor Vergata, Dept Syst Med, Chair Endocrinol & Med Sexol ENDOSEX, Rome, Italy
关键词
global healthcare; equity; artificial intelligence; ChatGPT; language bias;
D O I
10.3389/fmed.2023.1237432
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
BackgroundThe demand for healthcare is increasing globally, with notable disparities in access to resources, especially in Asia, Africa, and Latin America. The rapid development of Artificial Intelligence (AI) technologies, such as OpenAI's ChatGPT, has shown promise in revolutionizing healthcare. However, potential challenges, including the need for specialized medical training, privacy concerns, and language bias, require attention.MethodsTo assess the applicability and limitations of ChatGPT in Chinese and English settings, we designed an experiment evaluating its performance in the 2022 National Medical Licensing Examination (NMLE) in China. For a standardized evaluation, we used the comprehensive written part of the NMLE, translated into English by a bilingual expert. All questions were input into ChatGPT, which provided answers and reasons for choosing them. Responses were evaluated for "information quality" using the Likert scale.ResultsChatGPT demonstrated a correct response rate of 81.25% for Chinese and 86.25% for English questions. Logistic regression analysis showed that neither the difficulty nor the subject matter of the questions was a significant factor in AI errors. The Brier Scores, indicating predictive accuracy, were 0.19 for Chinese and 0.14 for English, indicating good predictive performance. The average quality score for English responses was excellent (4.43 point), slightly higher than for Chinese (4.34 point).ConclusionWhile AI language models like ChatGPT show promise for global healthcare, language bias is a key challenge. Ensuring that such technologies are robustly trained and sensitive to multiple languages and cultures is vital. Further research into AI's role in healthcare, particularly in areas with limited resources, is warranted.
引用
收藏
页数:7
相关论文
共 44 条
  • [1] The performance evaluation of artificial intelligence ERNIE bot in Chinese National Medical Licensing Examination
    Huang, Leiyun
    Hu, Jinghan
    Cai, Qingjin
    Fu, Guangjie
    Bai, Zhenglin
    Liu, Yongzhen
    Zheng, Ji
    Meng, Zengdong
    POSTGRADUATE MEDICAL JOURNAL, 2024, 100 (1190) : 952 - 953
  • [2] ChatGPT Performs on the Chinese National Medical Licensing Examination
    Xinyi Wang
    Zhenye Gong
    Guoxin Wang
    Jingdan Jia
    Ying Xu
    Jialu Zhao
    Qingye Fan
    Shaun Wu
    Weiguo Hu
    Xiaoyang Li
    Journal of Medical Systems, 47
  • [3] ChatGPT Performs on the Chinese National Medical Licensing Examination
    Wang, Xinyi
    Gong, Zhenye
    Wang, Guoxin
    Jia, Jingdan
    Xu, Ying
    Zhao, Jialu
    Fan, Qingye
    Wu, Shaun
    Hu, Weiguo
    Li, Xiaoyang
    JOURNAL OF MEDICAL SYSTEMS, 2023, 47 (01)
  • [4] The Performance of Artificial Intelligence on a National Medical Licensing Examination
    Geissler, Mark Enrik
    Goeben, Merle
    Glasmacher, Kira A.
    Bereuter, Jean-Paul
    Geissler, Rona Berit
    Wiest, Isabella C.
    Kolbinger, Fiona R.
    Kather, Jakob Nikolas
    DEUTSCHES ARZTEBLATT INTERNATIONAL, 2024, 121 (26): : 888 - 889
  • [5] Accuracy of ChatGPT on Medical Questions in the National Medical Licensing Examination in Japan: Evaluation Study
    Yanagita, Yasutaka
    Yokokawa, Daiki
    Uchida, Shun
    Tawara, Junsuke
    Ikusaka, Masatomi
    JMIR FORMATIVE RESEARCH, 2023, 7
  • [6] ChatGPT-4: An assessment of an upgraded artificial intelligence chatbot in the United States Medical Licensing Examination
    Mihalache, Andrew
    Huang, Ryan S.
    Popovic, Marko M.
    Muni, Rajeev H.
    MEDICAL TEACHER, 2024, 46 (03) : 366 - 372
  • [7] Performance of ChatGPT on Chinese national medical licensing examinations: a five-year examination evaluation study for physicians, pharmacists and nurses
    Zong, Hui
    Li, Jiakun
    Wu, Erman
    Wu, Rongrong
    Lu, Junyu
    Shen, Bairong
    BMC MEDICAL EDUCATION, 2024, 24 (01)
  • [8] Performance of ChatGPT on Chinese national medical licensing examinations: a five-year examination evaluation study for physicians, pharmacists and nurses
    Hui Zong
    Jiakun Li
    Erman Wu
    Rongrong Wu
    Junyu Lu
    Bairong Shen
    BMC Medical Education, 24
  • [9] Can ChatGPT pass China's national medical licensing examination?
    Shang, Luxiang
    Xue, Mingyue
    Hou, Yinglong
    Tang, Baopeng
    ASIAN JOURNAL OF SURGERY, 2023, 46 (12) : 6112 - 6113
  • [10] On the Relationship Between Open Science in Artificial Intelligence for Medical Imaging and Global Health Equity
    Souza, Raissa
    Stanley, Emma A. M.
    Forkert, Nils D.
    CLINICAL IMAGE-BASED PROCEDURES, FAIRNESS OF AI IN MEDICAL IMAGING, AND ETHICAL AND PHILOSOPHICAL ISSUES IN MEDICAL IMAGING, CLIP 2023, FAIMI 2023, EPIMI 2023, 2023, 14242 : 289 - 300