Is ChatGPT accurate and reliable in answering questions regarding head and neck cancer?

被引:38
|
作者
Kuscu, Oguz [1 ]
Pamuk, A. Erim [1 ]
Suslu, Nilda Sutay [2 ]
Hosal, Sefik [2 ]
机构
[1] Hacettepe Univ, Sch Med, Dept Otorhinolaryngol, TR-06100 Ankara, Turkiye
[2] Atılım Univ, Sch Med, Dept Otorhinolaryngol, Ankara, Turkiye
来源
FRONTIERS IN ONCOLOGY | 2023年 / 13卷
关键词
ChatGPT; 4; head and neck (H&N) cancer; head and neck; artificial intelligence; chatbot; information literacy; natural language processing; machine learning; MODELS;
D O I
10.3389/fonc.2023.1256459
中图分类号
R73 [肿瘤学];
学科分类号
100214 ;
摘要
Background and objective Chat Generative Pre-trained Transformer (ChatGPT) is an artificial intelligence (AI)-based language processing model using deep learning to create human-like text dialogue. It has been a popular source of information covering vast number of topics including medicine. Patient education in head and neck cancer (HNC) is crucial to enhance the understanding of patients about their medical condition, diagnosis, and treatment options. Therefore, this study aims to examine the accuracy and reliability of ChatGPT in answering questions regarding HNC.Methods 154 head and neck cancer-related questions were compiled from sources including professional societies, institutions, patient support groups, and social media. These questions were categorized into topics like basic knowledge, diagnosis, treatment, recovery, operative risks, complications, follow-up, and cancer prevention. ChatGPT was queried with each question, and two experienced head and neck surgeons assessed each response independently for accuracy and reproducibility. Responses were rated on a scale: (1) comprehensive/correct, (2) incomplete/partially correct, (3) a mix of accurate and inaccurate/misleading, and (4) completely inaccurate/irrelevant. Discrepancies in grading were resolved by a third reviewer. Reproducibility was evaluated by repeating questions and analyzing grading consistency.Results ChatGPT yielded "comprehensive/correct" responses to 133/154 (86.4%) of the questions whereas, rates of "incomplete/partially correct" and "mixed with accurate and inaccurate data/misleading" responses were 11% and 2.6%, respectively. There were no "completely inaccurate/irrelevant" responses. According to category, the model provided "comprehensive/correct" answers to 80.6% of questions regarding "basic knowledge", 92.6% related to "diagnosis", 88.9% related to "treatment", 80% related to "recovery - operative risks - complications - follow-up", 100% related to "cancer prevention" and 92.9% related to "other". There was not any significant difference between the categories regarding the grades of ChatGPT responses (p=0.88). The rate of reproducibility was 94.1% (145 of 154 questions).Conclusion ChatGPT generated substantially accurate and reproducible information to diverse medical queries related to HNC. Despite its limitations, it can be a useful source of information for both patients and medical professionals. With further developments in the model, ChatGPT can also play a crucial role in clinical decision support to provide the clinicians with up-to-date information.
引用
收藏
页数:7
相关论文
共 50 条
  • [1] Answering head and neck cancer questions: An assessment of ChatGPT responses
    Wei, Kimberly
    Fritz, Christian
    Rajasekaran, Karthik
    AMERICAN JOURNAL OF OTOLARYNGOLOGY, 2024, 45 (01)
  • [2] Assessing ChatGPT's performance in answering common questions regarding endometrial cancer
    Kuo, Iris
    Falk, Lauren
    Ladanyi, Andras
    Dewdney, Summer
    GYNECOLOGIC ONCOLOGY, 2024, 190 : S278 - S279
  • [3] Assessing the Knowledge of ChatGPT in Answering Questions Regarding Female Urology
    Cakir, Hakan
    Caglar, Ufuk
    Halis, Ahmet
    Sarilar, Omer
    Yazili, Huseyin Burak
    Ozgor, Faruk
    UROLOGY JOURNAL, 2024, 21 (06) : 410 - 414
  • [4] Assessing the performance of ChatGPT in answering questions regarding cirrhosis and hepatocellular carcinoma
    Yeo, Yee Hui
    Yang, Ju Dong
    CLINICAL AND MOLECULAR HEPATOLOGY, 2024, 30 (01)
  • [5] Letter 2 regarding "Assessing the performance of ChatGPT in answering questions regarding cirrhosis and hepatocellular carcinoma"
    Zhang, Yiwen
    Wu, Liwei
    Mu, Zepeng
    Ren, Linlin
    Chen, Ying
    Liu, Hanyun
    Xu, Lili
    Wang, Yangang
    Wang, Yaxing
    Cheng, Susan
    Tham, Yih Chung
    Sheng, Bin
    Wong, Tien Yin
    Ji, Hongwei
    CLINICAL AND MOLECULAR HEPATOLOGY, 2024, 30 (01)
  • [6] Letter 1 regarding "Assessing the performance of ChatGPT in answering questions regarding cirrhosis and hepatocellular carcinoma"
    Ali, Hassam
    CLINICAL AND MOLECULAR HEPATOLOGY, 2023, 29 (03) : 813 - 814
  • [7] Is ChatGPT reliable and accurate in answering pharmacotherapy-related inquiries in both Turkish and English?
    Ozturk, Nur
    Yakak, Irem
    Ag, Melih Bugra
    Aksoy, Nilay
    CURRENTS IN PHARMACY TEACHING AND LEARNING, 2024, 16 (07)
  • [8] Letter 2 regarding "Assessing the performance of ChatGPT in answering questions regarding cirrho- sis and hepatocellular carcinoma"
    Kleebayoon, Amnuay
    Wiwanitkit, Viroj
    CLINICAL AND MOLECULAR HEPATOLOGY, 2023, 29 (03) : 815 - 816
  • [9] Evaluating ChatGPT's performance in answering common patient questions on cervical cancer
    Do, Anthony
    Li, Andrew
    Smith, Haller
    Chambers, Laura
    Esselen, Kate
    Liang, Margaret
    GYNECOLOGIC ONCOLOGY, 2024, 190 : S376 - S376
  • [10] Evaluating the performance of ChatGPT in answering questions related to urolithiasis
    Hakan Cakir
    Ufuk Caglar
    Oguzhan Yildiz
    Arda Meric
    Ali Ayranci
    Faruk Ozgor
    International Urology and Nephrology, 2024, 56 : 17 - 21