Evaluation of the quality and readability of ChatGPT responses to frequently asked questions about myopia in traditional Chinese language.

IF 2.9 3区 医学 Q2 HEALTH CARE SCIENCES & SERVICES
DIGITAL HEALTH Pub Date : 2024-09-02 eCollection Date: 2024-01-01 DOI:10.1177/20552076241277021
Li-Chun Chang, Chi-Chin Sun, Ting-Han Chen, Der-Chong Tsai, Hui-Ling Lin, Li-Ling Liao
{"title":"Evaluation of the quality and readability of ChatGPT responses to frequently asked questions about myopia in traditional Chinese language.","authors":"Li-Chun Chang, Chi-Chin Sun, Ting-Han Chen, Der-Chong Tsai, Hui-Ling Lin, Li-Ling Liao","doi":"10.1177/20552076241277021","DOIUrl":null,"url":null,"abstract":"<p><strong>Introduction: </strong>ChatGPT can serve as an adjunct informational tool for ophthalmologists and their patients. However, the reliability and readability of its responses to myopia-related queries in the Chinese language remain underexplored.</p><p><strong>Purpose: </strong>This study aimed to evaluate the ability of ChatGPT to address frequently asked questions (FAQs) about myopia by parents and caregivers.</p><p><strong>Method: </strong>Myopia-related FAQs were input three times into fresh ChatGPT sessions, and the responses were evaluated by 10 ophthalmologists using a Likert scale for appropriateness, usability, and clarity. The Chinese Readability Index Explorer (CRIE) was used to evaluate the readability of each response. Inter-rater reliability among the reviewers was examined using Cohen's kappa coefficient, and Spearman's rank correlation analysis and one-way analysis of variance were used to investigate the relationship between CRIE scores and each criterion.</p><p><strong>Results: </strong>Forty-five percent of the responses of ChatGPT in Chinese language were appropriate and usable and only 35% met all the set criteria. The CRIE scores for 20 ChatGPT responses ranged from 7.29 to 12.09, indicating that the readability level was equivalent to a middle-to-high school level. Responses about the treatment efficacy and side effects were deficient for all three criteria.</p><p><strong>Conclusions: </strong>The performance of ChatGPT in addressing pediatric myopia-related questions is currently suboptimal. As parents increasingly utilize digital resources to obtain health information, it has become crucial for eye care professionals to familiarize themselves with artificial intelligence-driven information on pediatric myopia.</p>","PeriodicalId":51333,"journal":{"name":"DIGITAL HEALTH","volume":null,"pages":null},"PeriodicalIF":2.9000,"publicationDate":"2024-09-02","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11369861/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"DIGITAL HEALTH","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1177/20552076241277021","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2024/1/1 0:00:00","PubModel":"eCollection","JCR":"Q2","JCRName":"HEALTH CARE SCIENCES & SERVICES","Score":null,"Total":0}
引用次数: 0

Abstract

Introduction: ChatGPT can serve as an adjunct informational tool for ophthalmologists and their patients. However, the reliability and readability of its responses to myopia-related queries in the Chinese language remain underexplored.

Purpose: This study aimed to evaluate the ability of ChatGPT to address frequently asked questions (FAQs) about myopia by parents and caregivers.

Method: Myopia-related FAQs were input three times into fresh ChatGPT sessions, and the responses were evaluated by 10 ophthalmologists using a Likert scale for appropriateness, usability, and clarity. The Chinese Readability Index Explorer (CRIE) was used to evaluate the readability of each response. Inter-rater reliability among the reviewers was examined using Cohen's kappa coefficient, and Spearman's rank correlation analysis and one-way analysis of variance were used to investigate the relationship between CRIE scores and each criterion.

Results: Forty-five percent of the responses of ChatGPT in Chinese language were appropriate and usable and only 35% met all the set criteria. The CRIE scores for 20 ChatGPT responses ranged from 7.29 to 12.09, indicating that the readability level was equivalent to a middle-to-high school level. Responses about the treatment efficacy and side effects were deficient for all three criteria.

Conclusions: The performance of ChatGPT in addressing pediatric myopia-related questions is currently suboptimal. As parents increasingly utilize digital resources to obtain health information, it has become crucial for eye care professionals to familiarize themselves with artificial intelligence-driven information on pediatric myopia.

评估中文繁体近视常见问题聊天 GPT 回答的质量和可读性。
简介ChatGPT 可作为眼科医生及其患者的辅助信息工具。目的:本研究旨在评估 ChatGPT 解决家长和护理人员关于近视的常见问题(FAQs)的能力:方法:将与近视相关的常见问题三次输入到新鲜的 ChatGPT 会话中,由 10 位眼科医生使用李克特量表对回复的适当性、可用性和清晰度进行评估。中文可读性指数资源管理器(CRIE)用于评估每个回复的可读性。使用科恩卡帕系数(Cohen's kappa coefficient)检验了审稿人之间的可靠性,并使用斯皮尔曼等级相关分析和单因素方差分析来研究 CRIE 分数与各项标准之间的关系:45%的中文 ChatGPT 回答是恰当和可用的,只有 35%的回答符合所有设定的标准。20 个 ChatGPT 回答的 CRIE 分数从 7.29 到 12.09 不等,表明可读性水平相当于初中至高中水平。关于疗效和副作用的回答在所有三个标准中都存在不足:目前,ChatGPT 在解决小儿近视相关问题方面的表现并不理想。随着家长越来越多地利用数字资源获取健康信息,眼科专业人员熟悉人工智能驱动的儿童近视信息已变得至关重要。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
DIGITAL HEALTH
DIGITAL HEALTH Multiple-
CiteScore
2.90
自引率
7.70%
发文量
302
文献相关原料
公司名称 产品信息 采购帮参考价格
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信