Large language models: a new frontier in paediatric cataract patient education

被引:3
|
作者
Dihan, Qais [1 ,2 ]
Chauhan, Muhammad Z. [2 ]
Eleiwa, Taher K. [3 ]
Brown, Andrew D. [4 ]
Hassan, Amr K. [5 ]
Khodeiry, Mohamed M. [6 ]
Elsheikh, Reem H. [2 ]
Oke, Isdin [7 ]
Nihalani, Bharti R. [7 ]
VanderVeen, Deborah K. [7 ]
Sallam, Ahmed B. [2 ]
Elhusseiny, Abdelrahman M. [2 ,7 ]
机构
[1] Rosalind Franklin Univ Med & Sci, Chicago Med Sch, N Chicago, IL USA
[2] Univ Arkansas Med Sci, Dept Ophthalmol, Little Rock, AR 72205 USA
[3] Benha Univ, Dept Ophthalmol, Banha, Egypt
[4] Univ Arkansas Med Sci, Little Rock, AR USA
[5] South Valley Univ, Dept Ophthalmol, Qena, Egypt
[6] Univ Kentucky, Dept Ophthalmol, Lexington, KY USA
[7] Harvard Med Sch, Boston Childrens Hosp, Dept Ophthalmol, Boston, MA 02115 USA
关键词
Medical Education; Public health; Epidemiology; Child health (paediatrics); CHILDHOOD; READABILITY; INFORMATION; QUALITY; HEALTH;
D O I
10.1136/bjo-2024-325252
中图分类号
R77 [眼科学];
学科分类号
100212 ;
摘要
Background/aims This was a cross-sectional comparative study. We evaluated the ability of three large language models (LLMs) (ChatGPT-3.5, ChatGPT-4, and Google Bard) to generate novel patient education materials (PEMs) and improve the readability of existing PEMs on paediatric cataract. Methods We compared LLMs' responses to three prompts. Prompt A requested they write a handout on paediatric cataract that was 'easily understandable by an average American.' Prompt B modified prompt A and requested the handout be written at a 'sixth-grade reading level, using the Simple Measure of Gobbledygook (SMOG) readability formula.' Prompt C rewrote existing PEMs on paediatric cataract 'to a sixth-grade reading level using the SMOG readability formula'. Responses were compared on their quality (DISCERN; 1 (low quality) to 5 (high quality)), understandability and actionability (Patient Education Materials Assessment Tool (>= 70%: understandable, >= 70%: actionable)), accuracy (Likert misinformation; 1 (no misinformation) to 5 (high misinformation) and readability (SMOG, Flesch-Kincaid Grade Level (FKGL); grade level <7: highly readable). Results All LLM-generated responses were of high-quality (median DISCERN >= 4), understandability (>= 70%), and accuracy (Likert=1). All LLM-generated responses were not actionable (<70%). ChatGPT-3.5 and ChatGPT-4 prompt B responses were more readable than prompt A responses (p<0.001). ChatGPT-4 generated more readable responses (lower SMOG and FKGL scores; 5.59 +/- 0.5 and 4.31 +/- 0.7, respectively) than the other two LLMs (p<0.001) and consistently rewrote them to or below the specified sixth-grade reading level (SMOG: 5.14 +/- 0.3). Conclusion LLMs, particularly ChatGPT-4, proved valuable in generating high-quality, readable, accurate PEMs and in improving the readability of existing materials on paediatric cataract.
引用
收藏
页数:7
相关论文
共 50 条
  • [21] The Frontier of Data Erasure: A Survey on Machine Unlearning for Large Language Models
    Qu, Youyang
    Ding, Ming
    Sun, Nan
    Thilakarathna, Kanchana
    Zhu, Tianqing
    Niyato, Dusit
    COMPUTER, 2025, 58 (01) : 45 - 57
  • [22] Patient- and clinician-based evaluation of large language models for patient education in prostate cancer radiotherapy
    Trapp, Christian
    Schmidt-Hegemann, Nina
    Keilholz, Michael
    Brose, Sarah Frederike
    Marschner, Sebastian N.
    Schoenecker, Stephan
    Maier, Sebastian H.
    Dehelean, Diana-Coralia
    Rottler, Maya
    Konnerth, Dinah
    Belka, Claus
    Corradini, Stefanie
    Rogowski, Paul
    STRAHLENTHERAPIE UND ONKOLOGIE, 2025, 201 (03) : 333 - 342
  • [23] Evaluating large language models as patient education tools for inflammatory bowel disease: A comparative study
    Zhang, Yan
    Wan, Xiao-Han
    Kong, Qing-Zhou
    Liu, Han
    Liu, Jun
    Guo, Jing
    Yang, Xiao-Yun
    Zuo, Xiu-Li
    Li, Yan-Qing
    WORLD JOURNAL OF GASTROENTEROLOGY, 2025, 31 (06)
  • [24] Large-Language Models in Orthodontics: Assessing Reliability and Validity of ChatGPT in Pretreatment Patient Education
    Vassis, Stratos
    Powell, Harriet
    Petersen, Emma
    Barkmann, Asta
    Noeldeke, Beatrice
    Kristensen, Kasper D.
    Stoustrup, Peter
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2024, 16 (08)
  • [25] Tailoring glaucoma education using large language models: Addressing health disparities in patient comprehension
    Spina, Aidin C.
    Fereydouni, Pirooz
    Tang, Jordan N.
    Andalib, Saman
    Picton, Bryce G.
    Fox, Austin R.
    MEDICINE, 2025, 104 (02)
  • [26] Assessing state-of-the-art online large language models for patient education regarding prostatitis
    Zhang, Pengfei
    Wang, Hui
    Li, Pengfei
    Fu, Xianchun
    Yuan, Hang
    Ji, Hongwei
    Niu, Haitao
    PROSTATE, 2024, 84 (12): : 1173 - 1175
  • [27] Evaluation of Online Large Language Models for Patient Education Regarding Chronic Low Back Pain
    Liu, Xiang
    Li, Pengfei
    Fan, Xiaohua
    Wang, Chao
    PAIN PHYSICIAN, 2024, 27 (05) : E645 - E647
  • [28] USING 3D PRINTER TECHNOLOGY TO MANUFACTURE ANATOMIC MODELS FOR PATIENT EDUCATION: A NEW FRONTIER
    Dubey, Arbind
    Pathak, Alok
    Sharma, Ankur
    Harris, Chad
    Rickey, Daniel
    Sasaki, David
    Koul, Rashmi
    RADIOTHERAPY AND ONCOLOGY, 2016, 120 : S79 - S80
  • [29] Large language models challenge the future of higher education
    Milano, Silvia
    McGrane, Joshua A.
    Leonelli, Sabina
    NATURE MACHINE INTELLIGENCE, 2023, 5 (04) : 333 - 334
  • [30] Large language models challenge the future of higher education
    Silvia Milano
    Joshua A. McGrane
    Sabina Leonelli
    Nature Machine Intelligence, 2023, 5 : 333 - 334