Evaluation of responses to cardiac imaging questions by the artificial intelligence large language model ChatGPT

被引:6
|
作者
Monroe, Cynthia L. [1 ]
Abdelhafez, Yasser G. [2 ]
Atsina, Kwame [3 ]
Aman, Edris [3 ]
Nardo, Lorenzo [2 ]
Madani, Mohammad H. [2 ]
机构
[1] Calif Northstate Univ, Coll Med, 9700 W Taron Dr, Elk Grove, CA 95757 USA
[2] Univ Calif Davis, Med Ctr, Dept Radiol, 4860 Y St,Suite 3100, Sacramento, CA 95817 USA
[3] Univ Calif Davis, Med Ctr, Div Cardiovasc Med, 4860 Y St,Suite 0200, Sacramento, CA 95817 USA
关键词
Accuracy; Cardiac imaging; ChatGPT; Patient education; EXPERT CONSENSUS DOCUMENT; COMPUTED-TOMOGRAPHY SCCT; CORONARY-ARTERY-DISEASE; AMERICAN-COLLEGE; RADIOLOGY ACR; SOCIETY;
D O I
10.1016/j.clinimag.2024.110193
中图分类号
R8 [特种医学]; R445 [影像诊断学];
学科分类号
1002 ; 100207 ; 1009 ;
摘要
Purpose: To assess ChatGPT 's ability as a resource for educating patients on various aspects of cardiac imaging, including diagnosis, imaging modalities, indications, interpretation of radiology reports, and management. Methods: 30 questions were posed to ChatGPT-3.5 and ChatGPT-4 three times in three separate chat sessions. Responses were scored as correct, incorrect, or clinically misleading categories by three observers -two board certified cardiologists and one board certified radiologist with cardiac imaging subspecialization. Consistency of responses across the three sessions was also evaluated. Final categorization was based on majority vote between at least two of the three observers. Results: ChatGPT-3.5 answered seventeen of twenty eight questions correctly (61 %) by majority vote. Twenty one of twenty eight questions were answered correctly (75 %) by ChatGPT-4 by majority vote. Majority vote for correctness was not achieved for two questions. Twenty six of thirty questions were answered consistently by ChatGPT-3.5 (87 %). Twenty nine of thirty questions were answered consistently by ChatGPT-4 (97 %). ChatGPT-3.5 had both consistent and correct responses to seventeen of twenty eight questions (61 %). ChatGPT-4 had both consistent and correct responses to twenty of twenty eight questions (71 %). Conclusion: ChatGPT-4 had overall better performance than ChatGTP-3.5 when answering cardiac imaging questions with regard to correctness and consistency of responses. While both ChatGPT-3.5 and ChatGPT-4 answers over half of cardiac imaging questions correctly, inaccurate, clinically misleading and inconsistent responses suggest the need for further refinement before its application for educating patients about cardiac imaging.
引用
收藏
页数:8
相关论文
共 50 条
  • [31] The rise of artificial intelligence: addressing the impact of large language models such as ChatGPT on scientific publications
    Ang, Tiing Leong
    Choolani, Mahesh
    See, Kay Choong
    Poh, Kian Keong
    SINGAPORE MEDICAL JOURNAL, 2023, 64 (04) : 219 - 221
  • [32] Evaluation High-Quality of Information from ChatGPT (Artificial IntelligencedLarge Language Model) Artificial Intelligence on Shoulder Stabilization Surgery
    Hurley, Eoghan T.
    Crook, Bryan S.
    Lorentz, Samuel G.
    Danilkowicz, Richard M.
    Lau, Brian C.
    Taylor, Dean C.
    Dickens, Jonathan F.
    Anakwenze, Oke
    Klifto, Christopher S.
    ARTHROSCOPY-THE JOURNAL OF ARTHROSCOPIC AND RELATED SURGERY, 2024, 40 (03): : 726 - 731.e6
  • [33] Language-adaptive artificial intelligence: assessing CHATGPT'S answer to frequently asked questions on total hip arthroplasty questions
    Ibrahim, Muhammad Talal
    Khaskheli, Sarah Ashraf
    Shahzad, Hania
    Noordin, Shahryar
    JOURNAL OF THE PAKISTAN MEDICAL ASSOCIATION, 2024, 74 (04) : S161 - S164
  • [34] Evaluation of ChatGPT-4 responses to glaucoma patients' questions: Can artificial intelligence become a trusted advisor between doctor and patient?
    Gueler, Muzaffer Said
    Baydemir, Elif Ertan
    CLINICAL AND EXPERIMENTAL OPHTHALMOLOGY, 2024, 52 (09): : 1016 - 1019
  • [35] Correspondence on “Assessing the Accuracy of Responses by the Language Model ChatGPT to Questions Regarding Bariatric Surgery”
    Namria Ishaaq
    Shahab Saquib Sohail
    Obesity Surgery, 2023, 33 : 4159 - 4159
  • [36] Correspondence on "Assessing the Accuracy of Responses by the Language Model ChatGPT to Questions Regarding Bariatric Surgery"
    Ishaaq, Namria
    Sohail, Shahab Saquib
    OBESITY SURGERY, 2023, 33 (12) : 4159 - 4159
  • [37] Artificial Intelligence for Cardiac Imaging
    Engelhardt, Sandy
    Martin, Simon
    Bolanos, Carlos Rodrigo Rodriguez
    Pappas, Laura
    Koehler, Sven
    Nagel, Eike
    AKTUELLE KARDIOLOGIE, 2023, 12 (06) : 459 - 466
  • [38] Artificial intelligence and multiple sclerosis: ChatGPT model
    Altunisik, Erman
    MULTIPLE SCLEROSIS AND RELATED DISORDERS, 2023, 76
  • [39] Unlocking the Power of ChatGPT, Artificial Intelligence, and Large Language Models: Practical Suggestions for Radiation Oncologists
    Waters, Michael R.
    Aneja, Sanjay
    Hong, Julian C.
    PRACTICAL RADIATION ONCOLOGY, 2023, 13 (06) : E484 - E490
  • [40] Artificial Intelligence and Cerebrovascular Diseases: ChatGPT model
    Altunisik, Erman
    CEREBROVASCULAR DISEASES, 2024, 53 (03) : 354 - 358