可读性
误传
医学
阅读(过程)
等级
利克特量表
理解力
质量(理念)
病人教育
家庭医学
医学教育
数学教育
计算机科学
心理学
发展心理学
哲学
计算机安全
认识论
政治学
法学
程序设计语言
作者
Qais A. Dihan,Muhammad Z. Chauhan,Taher K. Eleiwa,Andrew D. Brown,Amr K. Hassan,Mohamed M. Khodeiry,Reem H. ElSheikh,Isdin Oke,Bharti R. Nihalani,Deborah K. VanderVeen,Ahmed B. Sallam,Abdelrahman M. Elhusseiny
标识
DOI:10.1136/bjo-2024-325252
摘要
Background/aims This was a cross-sectional comparative study. We evaluated the ability of three large language models (LLMs) (ChatGPT-3.5, ChatGPT-4, and Google Bard) to generate novel patient education materials (PEMs) and improve the readability of existing PEMs on paediatric cataract. Methods We compared LLMs’ responses to three prompts. Prompt A requested they write a handout on paediatric cataract that was ‘easily understandable by an average American.’ Prompt B modified prompt A and requested the handout be written at a ‘sixth-grade reading level, using the Simple Measure of Gobbledygook (SMOG) readability formula.’ Prompt C rewrote existing PEMs on paediatric cataract ‘to a sixth-grade reading level using the SMOG readability formula’. Responses were compared on their quality (DISCERN; 1 (low quality) to 5 (high quality)), understandability and actionability (Patient Education Materials Assessment Tool (≥70%: understandable, ≥70%: actionable)), accuracy (Likert misinformation; 1 (no misinformation) to 5 (high misinformation) and readability (SMOG, Flesch-Kincaid Grade Level (FKGL); grade level <7: highly readable). Results All LLM-generated responses were of high-quality (median DISCERN ≥4), understandability (≥70%), and accuracy (Likert=1). All LLM-generated responses were not actionable (<70%). ChatGPT-3.5 and ChatGPT-4 prompt B responses were more readable than prompt A responses (p<0.001). ChatGPT-4 generated more readable responses (lower SMOG and FKGL scores; 5.59±0.5 and 4.31±0.7, respectively) than the other two LLMs (p<0.001) and consistently rewrote them to or below the specified sixth-grade reading level (SMOG: 5.14±0.3). Conclusion LLMs, particularly ChatGPT-4, proved valuable in generating high-quality, readable, accurate PEMs and in improving the readability of existing materials on paediatric cataract.
科研通智能强力驱动
Strongly Powered by AbleSci AI