Readability, reliability and quality of responses generated by ChatGPT, gemini, and perplexity for the most frequently asked questions about pain

被引:0
|
作者
Ozduran, Erkan [1 ]
Akkoc, Ibrahim [2 ]
Buyukcoban, Sibel [3 ]
Erkin, Yueksel [4 ]
Hanci, Volkan [5 ]
机构
[1] Sivas Numune Hosp, Phys Med & Rehabil Pain Med, Sivas, Turkiye
[2] Univ Hlth Sci, Basaksehir Cam & Sakura City Hosp, Anesthesiol & Reanimat, Istanbul, Turkiye
[3] Dokuz Eylul Univ, Anesthesiol & Reanimat, Izmir, Turkiye
[4] Dokuz Eylul Univ, Anesthesiol & Reanimat, Pain Med, Izmir, Turkiye
[5] Dokuz Eylul Univ, Crit Care Med, Anesthesiol & Reanimat, Izmir, Turkiye
关键词
artificial intelligence; ChatGPT; Gemini; online medical information; pain; perplexity; INFORMATION; IMPROVE; CARE;
D O I
10.1097/MD.0000000000041780
中图分类号
R5 [内科学];
学科分类号
1002 ; 100201 ;
摘要
It is clear that artificial intelligence-based chatbots will be popular applications in the field of healthcare in the near future. It is known that more than 30% of the world's population suffers from chronic pain and individuals try to access the health information they need through online platforms before applying to the hospital. This study aimed to examine the readability, reliability and quality of the responses given by 3 different artificial intelligence chatbots (ChatGPT, Gemini and Perplexity) to frequently asked questions about pain. In this study, the 25 most frequently used keywords related to pain were determined using Google Trend and asked to every 3 artificial intelligence chatbots. The readability of the response texts was determined by Flesch Reading Ease Score (FRES), Simple Measure of Gobbledygook, Gunning Fog and Flesch-Kincaid Grade Level readability scoring. Reliability assessment was determined by the Journal of American Medical Association (JAMA), DISCERN scales. Global Quality Score and Ensuring Quality Information for Patients (EQIP) score were used in quality assessment. As a result of Google Trend search, the first 3 keywords were determined as "back pain," "stomach pain," and "chest pain." The readability of the answers given by all 3 artificial intelligence applications was determined to be higher than the recommended 6th grade readability level (P < .001). In the readability evaluation, the order from easy to difficult was determined as Google Gemini, ChatGPT and Perplexity. Higher GQS scores (P = .008) were detected in Gemini compared to other chatbots. Perplexity had higher JAMA, DISCERN and EQIP scores compared to other chatbots, respectively (P < .001, P < .001, P < .05). It has been determined that the answers given by ChatGPT, Gemini, and Perplexity to pain-related questions are difficult to read and their reliability and quality are low. It can be stated that these artificial intelligence chatbots cannot replace a comprehensive medical consultation. In artificial intelligence applications, it may be recommended to facilitate the readability of text content, create texts containing reliable references, and control them by a supervisory expert team.
引用
收藏
页数:10
相关论文
共 50 条
  • [1] Evaluation of the quality and readability of ChatGPT responses to frequently asked questions about myopia in traditional Chinese language
    Chang, Li-Chun
    Sun, Chi-Chin
    Chen, Ting-Han
    Tsai, Der-Chong
    Lin, Hui-Ling
    Liao, Li-Ling
    DIGITAL HEALTH, 2024, 10
  • [2] Assessment of readability, reliability, and quality of ChatGPT®, BARD®, Gemini®, Copilot®, Perplexity® responses on palliative care
    Hanci, Volkan
    Ergun, Bisar
    Gul, Sanser
    Uzun, Ozcan
    Erdemir, Ismail
    Hanci, Ferid Baran
    MEDICINE, 2024, 103 (33)
  • [3] Acceptability and readability of ChatGPT-4 based responses for frequently asked questions about strabismus and amblyopia
    Guven, S.
    Ayyildiz, B.
    JOURNAL FRANCAIS D OPHTALMOLOGIE, 2025, 48 (03):
  • [4] Evaluation of information accuracy and clarity: ChatGPT responses to the most frequently asked questions about premature ejaculation
    Sahin, Mehmet Fatih
    Keles, Anil
    Ozcan, Ridvan
    Dogan, Cagri
    Topkac, Erdem Can
    Akgul, Murat
    Yazici, Cenk Murat
    SEXUAL MEDICINE, 2024, 12 (03)
  • [5] Responses to the 10 Most Frequently Asked Questions About Perfusion CT
    Leiva-Salinas, Carlos
    Provenzale, James M.
    Wintermark, Max
    AMERICAN JOURNAL OF ROENTGENOLOGY, 2011, 196 (01) : 53 - 60
  • [6] Assessing the Quality and Readability of Chatgpt Responses to Commonly Asked Questions in Plastic Surgery
    Keating, Muireann
    Bollard, Stephanie
    Potter, Shirley
    IRISH JOURNAL OF MEDICAL SCIENCE, 2024, 193 : S54 - S54
  • [7] Quality of ChatGPT Responses to Frequently Asked Questions in Carpal Tunnel Release Surgery
    Amen, Troy B.
    Torabian, Kaveh A.
    Subramanian, Tejas
    Yang, Brian W.
    Liimakka, Adriana
    Fufa, Duretti
    PLASTIC AND RECONSTRUCTIVE SURGERY-GLOBAL OPEN, 2024, 12 (05) : E5822
  • [8] An assessment of ChatGPT's responses to frequently asked questions about cervical and breast cancer
    Ye, Zichen
    Zhang, Bo
    Zhang, Kun
    Mendez, Maria Jose Gonzalez
    Yan, Huijiao
    Wu, Tong
    Qu, Yimin
    Jiang, Yu
    Xue, Peng
    Qiao, Youlin
    BMC WOMENS HEALTH, 2024, 24 (01)
  • [9] ChatGPT's response to frequently asked questions about ultrasonography
    Hong, Daorong
    Huang, Chunyan
    MEDICAL ULTRASONOGRAPHY, 2023, 25 (04) : 473 - 474
  • [10] Assessing the Responses of Large Language Models (ChatGPT-4, Gemini, and Microsoft Copilot) to Frequently Asked Questions in Breast Imaging: A Study on Readability and Accuracy
    Tepe, Murat
    Emekli, Emre
    CUREUS JOURNAL OF MEDICAL SCIENCE, 2024, 16 (05)