Appropriateness of ChatGPT in Answering Heart Failure Related Questions

被引:12
|
作者
King, Ryan C. [1 ]
Samaan, Jamil S. [2 ]
Yeo, Yee Hui [2 ]
Mody, Behram [1 ]
Lombardo, Dawn M. [1 ]
Ghashghaei, Roxana [1 ]
机构
[1] Univ Calif Irvine, Irvine Med Ctr, Dept Med, Div Cardiol, 101 City Dr South, Orange, CA 92868 USA
[2] Cedars Sinai Med Ctr, Dept Med, Karsh Div Gastroenterol & Hepatol, Los Angeles, CA USA
来源
HEART LUNG AND CIRCULATION | 2024年 / 33卷 / 09期
关键词
Heart failure; ChatGPT; Health education; Artificial fi cial intelligence; Equity;
D O I
10.1016/j.hlc.2024.03.005
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
Background Heart failure requires complex management, and increased patient knowledge has been shown to improve outcomes. This study assessed the knowledge of Chat Generative Pre-trained Transformer (ChatGPT) and its appropriateness as a supplemental resource of information for patients with heart failure. Method A total of 107 frequently asked heart failure-related questions were included in 3 categories: "basic knowledge" (49), "management" (41) and "other" (17). Two responses per question were generated using both GPT-3.5 and GPT-4 (i.e., two responses per question per model). The accuracy and reproducibility of responses were graded by two reviewers, board-certified fi ed in cardiology, with differences resolved by a third reviewer, board-certified fi ed in cardiology and advanced heart failure. Accuracy was graded using a four-point scale: (1) comprehensive, (2) correct but inadequate, (3) some correct and some incorrect, and (4) completely incorrect. Results GPT-4 provided 107/107 (100%) responses with correct information. Further, GPT-4 displayed a greater proportion of comprehensive knowledge for the categories of "basic knowledge" and "management" (89.8% and 82.9%, respectively). For GPT-3, there were two total responses (1.9%) graded as "some correct and incorrect" for GPT-3.5, while no "completely incorrect" responses were produced. With respect to comprehensive knowledge, GPT-3.5 performed best in the "management" category and "other" category (prognosis, procedures, and support) (78.1%, 94.1%). The models also provided highly reproducible responses, with GPT-3.5 scoring above 94% in every category and GPT-4 with 100% for all answers. Conclusions GPT-3.5 and GPT-4 answered the majority of heart failure-related questions accurately and reliably. If validated in future studies, ChatGPT may serve as a useful tool in the future by providing accessible health-related information and education to patients living with heart failure. In its current state, ChatGPT necessitates further rigorous testing and validation to ensure patient safety and equity across all patient demographics.
引用
收藏
页码:1314 / 1318
页数:5
相关论文
共 50 条
  • [21] Evaluation of ChatGPT and Gemini in Answering Patient Questions After Gynecologic Surgery
    Voigt, P.
    Sharma, R.
    Milad, M. P.
    Chaudhari, A.
    Tsai, S.
    Yang, L.
    OBSTETRICS AND GYNECOLOGY, 2025, 145 (5S): : 39S - 39S
  • [22] Is ChatGPT accurate and reliable in answering questions regarding head and neck cancer?
    Kuscu, Oguz
    Pamuk, A. Erim
    Suslu, Nilda Sutay
    Hosal, Sefik
    FRONTIERS IN ONCOLOGY, 2023, 13
  • [23] Assessing the performance of chat generative pretrained transformer (ChatGPT) in answering chronic kidney disease-related questions
    Can, Basak
    Kahvecioglu, Esra Deniz
    Palit, Fatih
    Cebeci, Egemen
    Kucuk, Mehmet
    Karaali, Zeynep
    THERAPEUTIC APHERESIS AND DIALYSIS, 2024,
  • [25] ChatGPT and Patients With Heart Failure
    Dimitriadis, Fotis
    Alkagiet, Stelina
    Tsigkriki, Lamprini
    Kleitsioti, Panagiota
    Sidiropoulos, George
    Efstratiou, Dimitris
    Askalidi, Taisa
    Tsaousidis, Adam
    Siarkos, Michail
    Giannakopoulou, Pinelopi
    Mavrogianni, Angeliki-Despoina
    Zarifis, John
    Koulaouzidis, George
    ANGIOLOGY, 2024,
  • [26] ChatGPT and Heart Failure: Comment
    Daungsupawong, Hinpetch
    Wiwanitkit, Viroj
    ANGIOLOGY, 2024,
  • [27] ChatGPT in Answering Queries Related to Lifestyle-Related Diseases and Disorders
    Mondal, Himel
    Dash, Ipsita
    Mondal, Shaikat
    Behera, Joshil Kumar
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2023, 15 (11)
  • [28] Assessing ChatGPT's performance in answering common questions regarding endometrial cancer
    Kuo, Iris
    Falk, Lauren
    Ladanyi, Andras
    Dewdney, Summer
    GYNECOLOGIC ONCOLOGY, 2024, 190 : S278 - S279
  • [29] Evaluation of the Accuracy of ChatGPT in Answering Clinical Questions on the Japanese Society of Hypertension Guidelines
    Kusunose, Kenya
    Kashima, Shuichiro
    Sata, Masataka
    CIRCULATION JOURNAL, 2023, 87 (07) : 1030 - 1033
  • [30] Evaluating ChatGPT's performance in answering common patient questions on cervical cancer
    Do, Anthony
    Li, Andrew
    Smith, Haller
    Chambers, Laura
    Esselen, Kate
    Liang, Margaret
    GYNECOLOGIC ONCOLOGY, 2024, 190 : S376 - S376