Performance of ChatGPT in Solving Questions From the Progress Test (Brazilian National Medical Exam): A Potential Artificial Intelligence Tool in Medical Practice

被引:1
|
作者
Alessi, Mateus Rodrigues [1 ]
Gomes, Heitor A. [1 ]
de Castro, Matheus Lopes [1 ]
Okamoto, Cristina Terumy [2 ]
机构
[1] Univ Posit, Sch Med, Curitiba, Brazil
[2] Univ Posit, Neonatol, Curitiba, Brazil
关键词
academic performance/grades; biomedical technology; medical codes of ethics; examination questions; expert systems; artificial intelligence and education; artificial intelligence in medicine;
D O I
10.7759/cureus.64924
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
Background The use of artificial intelligence (AI) is not a recent phenomenon, but the latest advancements in this technology are making a significant impact across various fields of human knowledge. In medicine, this trend is no different, although it has developed at a slower pace. ChatGPT is an example of an AI-based algorithm capable of answering questions, interpreting phrases, and synthesizing complex information, potentially aiding and even replacing humans in various areas of social interest. Some studies have compared its performance in solving medical knowledge exams with medical students and professionals to verify AI accuracy. This study aimed to measure the performance of ChatGPT in answering questions from the Progress Test from 2021 to 2023. Methodology An observational study was conducted in which questions from the 2021 Progress Test and the regional tests (Southern Institutional Pedagogical Support Center II) of 2022 and 2023 were presented to ChatGPT 3.5. The results obtained were compared with the scores of first- to sixth-year medical students from over 120 Brazilian universities. All questions were presented sequentially, without any modification to their structure. After each question was presented, the platform's history was cleared, and the site was restarted. Results The platform achieved an average accuracy rate in 2021, 2022, and 2023 of 69.7%, 68.3%, and 67.2%, respectively, surpassing students from all medical years in the three tests evaluated, reinforcing findings in the current literature. The subject with the best score for the AI was Public Health, with a mean grade of 77.8%. Conclusions ChatGPT demonstrated the ability to answer medical questions with higher accuracy than humans, including students from the last year of medical school.
引用
收藏
页数:10
相关论文
共 49 条
  • [1] Advancing Medical Practice with Artificial Intelligence: ChatGPT in Healthcare
    Tessler, Idit
    Wolfovitz, Amit
    Livneh, Nir
    Gecel, Nir A.
    Sorin, Vera
    Barash, Yiftach
    Konen, Eli
    Klang, Eyal
    ISRAEL MEDICAL ASSOCIATION JOURNAL, 2024, 26 (02): : 80 - 85
  • [2] Performance of ChatGPT-4 in answering questions from the Brazilian National Examination for Medical Degree Revalidation
    Wiwanitkit, Somsri
    Wiwanitkit, Viroj
    REVISTA DA ASSOCIACAO MEDICA BRASILEIRA, 2024, 70 (03):
  • [3] Performance of ChatGPT-4 in answering questions from the Brazilian National Examination for Medical Degree Revalidation
    Gobira, Mauro
    Nakayama, Luis Filipe
    Moreira, Rodrigo
    Andrade, Eric
    Regatieri, Caio Vinicius Saito
    Belfort Jr, Rubens
    REVISTA DA ASSOCIACAO MEDICA BRASILEIRA, 2023, 69 (10):
  • [4] ChatGPT Conquers the Saudi Medical Licensing Exam: Exploring the Accuracy of Artificial Intelligence in Medical Knowledge Assessment and Implications for Modern Medical Education
    Aljindan, Fahad K.
    Al Qurashi, Abdullah A.
    Albalawi, Ibrahim Abdullah S.
    Alanazi, Abeer Mohammed M.
    Aljuhani, Hussam Abdulkhaliq M.
    Almutairi, Faisal Falah
    Aldamigh, Omar A.
    Halawani, Ibrahim R.
    Alarki, Subhi M. K. Zino
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2023, 15 (09)
  • [5] Potential and limitations of ChatGPT and generative artificial intelligence in medical safety education
    Wang, Xin
    Liu, Xin-Qiao
    WORLD JOURNAL OF CLINICAL CASES, 2023, 11 (32)
  • [6] ChatGPT and Generative Artificial Intelligence for Medical Education: Potential Impact and Opportunity
    Boscardin, Christy K.
    Gin, Brian
    Golde, Polo Black
    Hauer, Karen E.
    ACADEMIC MEDICINE, 2024, 99 (01) : 22 - 27
  • [7] Is "ChatGPT" capable of passing the 2022 MIR exam? Implications of artificial intelligence in medical education in Spain
    Carrasco, J. P.
    Garcia, E.
    Sanchez, D. A.
    Porter, P. D. Estrella
    De La Puente, L.
    Navarro, J.
    Cerame, A.
    SPANISH JOURNAL OF MEDICAL EDUCATION, 2023, 4 (01):
  • [8] The Performance of Artificial Intelligence on a National Medical Licensing Examination
    Geissler, Mark Enrik
    Goeben, Merle
    Glasmacher, Kira A.
    Bereuter, Jean-Paul
    Geissler, Rona Berit
    Wiest, Isabella C.
    Kolbinger, Fiona R.
    Kather, Jakob Nikolas
    DEUTSCHES ARZTEBLATT INTERNATIONAL, 2024, 121 (26): : 888 - 889
  • [9] A Comparative Analysis of ChatGPT and Medical Faculty Graduates in Medical Specialization Exams: Uncovering the Potential of Artificial Intelligence in Medical Education
    Gencer, Gulcan
    Gencer, Kerem
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2024, 16 (08)
  • [10] Comparing the performance of artificial intelligence learning models to medical students in solving histology and embryology multiple choice questions
    Bajcetic, Milos
    Mircic, Aleksandar
    Rakocevic, Jelena
    Dokovic, Danilo
    Milutinovic, Katarina
    Zaletel, Ivan
    ANNALS OF ANATOMY-ANATOMISCHER ANZEIGER, 2024, 254