Using Large Language Models to Generate Educational Materials on Childhood Glaucoma

被引:3
|
作者
Dihan, Qais [1 ,2 ]
Chauhan, Muhammad z. [2 ]
Eleiwa, Taher k. [3 ]
Hassan, Amr k. [4 ]
Sallam, Ahmed b. [2 ,5 ]
Khouri, Albert s. [6 ]
Chang, Ta c. [7 ]
Elhusseiny, Abdelrahman m. [2 ,8 ]
机构
[1] Chicago Med Sch, Dept Med, N Chicago, IL USA
[2] Univ Arkansas Med Sci, Harvey & Bernice Jones Eye Inst, Dept Ophthalmol, Little Rock, AR USA
[3] Univ Arkansas Med Sci, Harvey & Bernice Jones Eye Inst, Benha, AR USA
[4] South Valley Univ, Fac Med, Dept Ophthalmol, Qena, Egypt
[5] Ain Shams Univ, Fac Med, Dept Ophthalmol, Cairo, Egypt
[6] Rutgers New Jersey Med Sch, Inst Ophthalmol & Visual Sci ASK, Newark, NJ USA
[7] Univ Miami, Bascom Palmer Eye Inst, Dept Ophthalmol, Miller Sch Med, Miami, FL USA
[8] Harvard Med Sch, Boston Childrens Hosp, Dept Ophthalmol, Boston, MA USA
关键词
FOLLOW-UP; READABILITY; INFORMATION; ADHERENCE; BARRIERS; QUALITY; CARE;
D O I
10.1016/j.ajo.2024.04.004
中图分类号
R77 [眼科学];
学科分类号
100212 ;
摘要
Purpose: To evaluate the quality, readability, and accuracy of large language model (LLM)-generated patient education materials (PEMs) on childhood glaucoma, and their ability to improve existing the readability of online information. Design: Cross-sectional comparative study. Methods: We evaluated responses of ChatGPT-3.5, ChatGPT-4, and Bard to 3 separate prompts requesting that they write PEMs on "childhood glaucoma." Prompt A required PEMs be "easily understandable by the average American." Prompt B required that PEMs be written "at a 6th-grade level using Simple Measure of Gobbledygook (SMOG) readability formula." We then compared responses' quality (DISCERN questionnaire, Patient Education Materials Assessment Tool [PEMAT]), readability (SMOG, Flesch-Kincaid Grade Level [FKGL]), and accuracy (Likert Misinformation scale). To assess the improvement of readability for existing online information, Prompt C requested that LLM rewrite 20 resources from a Google search of keyword "childhood glaucoma" to the American Medical Association-recommended "6th-grade level." Rewrites were compared on key metrics such as readability, complex words (>= 3 syllables), and sentence count. Results: All 3 LLMs generated PEMs that were of high quality, understandability, and accuracy (DISCERN >= 4, >= 70% PEMAT understandability, Misinformation score = 1). Prompt B responses were more readable than Prompt A responses for all 3 LLM (P <= .001). ChatGPT-4 generated the most readable PEMs compared to ChatGPT-3.5 and Bard (P <= .001). Although Prompt C responses showed consistent reduction of mean SMOG and FKGL scores, only ChatGPT-4 achieved the specified 6th-grade reading level (4.8 +/- 0.8 and 3.7 +/- 1.9, respectively). Conclusions:<bold> </bold>LLMs can serve as strong supplemental tools in generating high-quality, accurate, and novel PEMs, and improving the readability of existing PEMs on childhood glaucoma.
引用
收藏
页码:28 / 38
页数:11
相关论文
共 50 条
  • [41] Pilot Study on Using Large Language Models for Educational Resource Development in Japanese Radiological Technologist Exams
    Kondo, Tatsuya
    Okamoto, Masashi
    Kondo, Yohan
    MEDICAL SCIENCE EDUCATOR, 2025,
  • [42] MaScQA: investigating materials science knowledge of large language models
    Zaki, Mohd
    Jayadeva
    Mausam
    Krishnan, N. M. Anoop
    DIGITAL DISCOVERY, 2024, 3 (02): : 313 - 327
  • [43] Using Large Language Models in Business Processes
    Grisold, Thomas
    vom Brocke, Jan
    Kratsch, Wolfgang
    Mendling, Jan
    Vidgof, Maxim
    BUSINESS PROCESS MANAGEMENT, BPM 2023, 2023, 14159 : XXIX - XXXI
  • [44] Accelerating Pharmacovigilance using Large Language Models
    Prakash, Mukkamala Venkata Sai
    Parab, Ganesh
    Veeramalla, Meghana
    Reddy, Siddartha
    Varun, V.
    Gopalakrishnan, Saisubramaniam
    Pagidipally, Vishal
    Vaddina, Vishal
    PROCEEDINGS OF THE 17TH ACM INTERNATIONAL CONFERENCE ON WEB SEARCH AND DATA MINING, WSDM 2024, 2024, : 1182 - 1183
  • [45] Predicting Glaucoma Before Onset Using a Large Language Model Chatbot
    Huang, Xiaoqin
    Raja, Hina
    Madadi, Yeganeh
    Delsoz, Mohammad
    Poursoroush, Asma
    Kahook, Malik Y.
    Yousefi, Siamak
    AMERICAN JOURNAL OF OPHTHALMOLOGY, 2024, 266 : 289 - 299
  • [46] Predicting glaucoma before onset using a large language model chatbot
    Yousefi, Siamak
    Huang, Xiaoqin
    Raja, Hina
    Madadi, Yeganeh
    Delsoz, Mohammad
    Poursoroush, Asma
    Kahook, Malik
    INVESTIGATIVE OPHTHALMOLOGY & VISUAL SCIENCE, 2024, 65 (07)
  • [47] Leveraging Large Language Models to Generate Course-Specific Semantically Annotated Learning Objects
    Lohr, Dominic
    Berges, Marc
    Chugh, Abhishek
    Kohlhase, Michael
    Mueller, Dennis
    JOURNAL OF COMPUTER ASSISTED LEARNING, 2025, 41 (01)
  • [48] Intergenerational Learning as an Opportunity to Generate New Educational Models
    Cabanillas, Carmen
    JOURNAL OF INTERGENERATIONAL RELATIONSHIPS, 2011, 9 (02) : 229 - 231
  • [49] Towards Human-Like Educational Question Generation with Large Language Models
    Wang, Zichao
    Valdez, Jakob
    Mallick, Debshila Basu
    Baraniuk, Richard G.
    ARTIFICIAL INTELLIGENCE IN EDUCATION, PT I, 2022, 13355 : 153 - 166
  • [50] Transforming Educational Assessment: Insights Into the Use of ChatGPT and Large Language Models in Grading
    Kooli, Chokri
    Yusuf, Nadia
    INTERNATIONAL JOURNAL OF HUMAN-COMPUTER INTERACTION, 2025, 41 (05) : 3388 - 3399