Appropriateness of ChatGPT in Answering Heart Failure Related Questions

被引:12
|
作者
King, Ryan C. [1 ]
Samaan, Jamil S. [2 ]
Yeo, Yee Hui [2 ]
Mody, Behram [1 ]
Lombardo, Dawn M. [1 ]
Ghashghaei, Roxana [1 ]
机构
[1] Univ Calif Irvine, Irvine Med Ctr, Dept Med, Div Cardiol, 101 City Dr South, Orange, CA 92868 USA
[2] Cedars Sinai Med Ctr, Dept Med, Karsh Div Gastroenterol & Hepatol, Los Angeles, CA USA
来源
HEART LUNG AND CIRCULATION | 2024年 / 33卷 / 09期
关键词
Heart failure; ChatGPT; Health education; Artificial fi cial intelligence; Equity;
D O I
10.1016/j.hlc.2024.03.005
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
Background Heart failure requires complex management, and increased patient knowledge has been shown to improve outcomes. This study assessed the knowledge of Chat Generative Pre-trained Transformer (ChatGPT) and its appropriateness as a supplemental resource of information for patients with heart failure. Method A total of 107 frequently asked heart failure-related questions were included in 3 categories: "basic knowledge" (49), "management" (41) and "other" (17). Two responses per question were generated using both GPT-3.5 and GPT-4 (i.e., two responses per question per model). The accuracy and reproducibility of responses were graded by two reviewers, board-certified fi ed in cardiology, with differences resolved by a third reviewer, board-certified fi ed in cardiology and advanced heart failure. Accuracy was graded using a four-point scale: (1) comprehensive, (2) correct but inadequate, (3) some correct and some incorrect, and (4) completely incorrect. Results GPT-4 provided 107/107 (100%) responses with correct information. Further, GPT-4 displayed a greater proportion of comprehensive knowledge for the categories of "basic knowledge" and "management" (89.8% and 82.9%, respectively). For GPT-3, there were two total responses (1.9%) graded as "some correct and incorrect" for GPT-3.5, while no "completely incorrect" responses were produced. With respect to comprehensive knowledge, GPT-3.5 performed best in the "management" category and "other" category (prognosis, procedures, and support) (78.1%, 94.1%). The models also provided highly reproducible responses, with GPT-3.5 scoring above 94% in every category and GPT-4 with 100% for all answers. Conclusions GPT-3.5 and GPT-4 answered the majority of heart failure-related questions accurately and reliably. If validated in future studies, ChatGPT may serve as a useful tool in the future by providing accessible health-related information and education to patients living with heart failure. In its current state, ChatGPT necessitates further rigorous testing and validation to ensure patient safety and equity across all patient demographics.
引用
收藏
页码:1314 / 1318
页数:5
相关论文
共 50 条
  • [1] Assessing the applicability and appropriateness of ChatGPT in answering clinical pharmacy questions
    Fournier, A.
    Fallet, C.
    Sadeghipour, F.
    Perrottet, N.
    ANNALES PHARMACEUTIQUES FRANCAISES, 2024, 82 (03): : 507 - 513
  • [2] Appropriateness of ChatGPT as a resource for medication-related questions
    Grossman, Sara
    Zerilli, Tina
    Nathan, Joseph P.
    BRITISH JOURNAL OF CLINICAL PHARMACOLOGY, 2024, 90 (10) : 2691 - 2695
  • [3] Evaluating the performance of ChatGPT in answering questions related to urolithiasis
    Hakan Cakir
    Ufuk Caglar
    Oguzhan Yildiz
    Arda Meric
    Ali Ayranci
    Faruk Ozgor
    International Urology and Nephrology, 2024, 56 : 17 - 21
  • [4] Evaluating the performance of ChatGPT in answering questions related to urolithiasis
    Cakir, Hakan
    Caglar, Ufuk
    Yildiz, Oguzhan
    Meric, Arda
    Ayranci, Ali
    Ozgor, Faruk
    INTERNATIONAL UROLOGY AND NEPHROLOGY, 2024, 56 (01) : 17 - 21
  • [5] ChatGPT in answering questions related to pediatric urology: Comment
    Kleebayoon, Amnuay
    Wiwanitkit, Viroj
    JOURNAL OF PEDIATRIC UROLOGY, 2024, 20 (01) : 28 - 28
  • [6] Evaluating the performance of ChatGPT in answering questions related to pediatric urology
    Caglar, Ufuk
    Yildiz, Oguzhan
    Meric, Arda
    Ayranci, Ali
    Gelmis, Mucahit
    Sarilar, Omer
    Ozgor, Faruk
    JOURNAL OF PEDIATRIC UROLOGY, 2024, 20 (01) : 26.e1 - 26.e5
  • [7] Evaluating the Performance of ChatGPT in answering questions related to benign prostate hyperplasia and prostate cancer
    Caglar, Ufuk
    Yildiz, Oguzhan
    Meric, Arda
    Ayranci, Ali
    Yusuf, Resit
    Sarilar, Omer
    Ozgor, Faruk
    MINERVA UROLOGY AND NEPHROLOGY, 2023, 75 (06): : 729 - 733
  • [8] Assessing the knowledge of ChatGPT and Google Gemini in answering peripheral artery disease-related questions
    Cetin, Hakki Kursat
    Demir, Tolga
    VASCULAR, 2025,
  • [9] Assessing the Performance of Chat Generative Pretrained Transformer (ChatGPT) in Answering Andrology-Related Questions
    Caglar, Ufuk
    Yildiz, Oguzhan
    Ozervarli, M. Firat
    Aydin, Resat
    Sarilar, Omer
    Ozgor, Faruk
    Ortac, Mazhar
    UROLOGY RESEARCH AND PRACTICE, 2023, 49 (06) : 365 - 369
  • [10] Response to letter to the editor re "Evaluating the performance of ChatGPT in answering questions related to pediatric urology"
    Caglar, Ufuk
    Ozgor, Faruk
    JOURNAL OF PEDIATRIC UROLOGY, 2024, 20 (01) : 27 - 27