Assessing ChatGPT's accuracy and reliability in asthma general knowledge: implications for artificial intelligence use in public health education.

IF 1.7 4区 医学 Q3 ALLERGY
Muhammad Thesa Ghozali
{"title":"Assessing ChatGPT's accuracy and reliability in asthma general knowledge: implications for artificial intelligence use in public health education.","authors":"Muhammad Thesa Ghozali","doi":"10.1080/02770903.2025.2450482","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Integrating Artificial Intelligence (AI) into public health education represents a pivotal advancement in medical knowledge dissemination, particularly for chronic diseases such as asthma. This study assesses the accuracy and comprehensiveness of ChatGPT, a conversational AI model, in providing asthma-related information.</p><p><strong>Methods: </strong>Employing a rigorous mixed-methods approach, healthcare professionals evaluated ChatGPT's responses to the Asthma General Knowledge Questionnaire for Adults (AGKQA), a standardized instrument covering various asthma-related topics. Responses were graded for accuracy and completeness and analyzed using statistical tests to assess reproducibility and consistency.</p><p><strong>Results: </strong>ChatGPT showed notable proficiency in conveying asthma knowledge, with flawless success in the etiology and pathophysiology categories and substantial accuracy in medication information (70%). However, limitations were noted in medication-related responses, where mixed accuracy (30%) highlights the need for further refinement of ChatGPT's capabilities to ensure reliability in critical areas of asthma education. Reproducibility analysis demonstrated a consistent 100% rate across all categories, affirming ChatGPT's reliability in delivering uniform information. Statistical analyses further underscored ChatGPT's stability and reliability.</p><p><strong>Conclusion: </strong>These findings underscore ChatGPT's promise as a valuable educational tool for asthma while emphasizing the necessity of ongoing improvements to address observed limitations, particularly regarding medication-related information.</p>","PeriodicalId":15076,"journal":{"name":"Journal of Asthma","volume":" ","pages":"1-9"},"PeriodicalIF":1.7000,"publicationDate":"2025-01-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Asthma","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.1080/02770903.2025.2450482","RegionNum":4,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"ALLERGY","Score":null,"Total":0}
引用次数: 0

Abstract

Background: Integrating Artificial Intelligence (AI) into public health education represents a pivotal advancement in medical knowledge dissemination, particularly for chronic diseases such as asthma. This study assesses the accuracy and comprehensiveness of ChatGPT, a conversational AI model, in providing asthma-related information.

Methods: Employing a rigorous mixed-methods approach, healthcare professionals evaluated ChatGPT's responses to the Asthma General Knowledge Questionnaire for Adults (AGKQA), a standardized instrument covering various asthma-related topics. Responses were graded for accuracy and completeness and analyzed using statistical tests to assess reproducibility and consistency.

Results: ChatGPT showed notable proficiency in conveying asthma knowledge, with flawless success in the etiology and pathophysiology categories and substantial accuracy in medication information (70%). However, limitations were noted in medication-related responses, where mixed accuracy (30%) highlights the need for further refinement of ChatGPT's capabilities to ensure reliability in critical areas of asthma education. Reproducibility analysis demonstrated a consistent 100% rate across all categories, affirming ChatGPT's reliability in delivering uniform information. Statistical analyses further underscored ChatGPT's stability and reliability.

Conclusion: These findings underscore ChatGPT's promise as a valuable educational tool for asthma while emphasizing the necessity of ongoing improvements to address observed limitations, particularly regarding medication-related information.

评估ChatGPT在哮喘常识中的准确性和可靠性:人工智能在公共卫生教育中的应用
将人工智能(AI)纳入公共卫生教育是医学知识传播的关键进步,特别是对于哮喘等慢性疾病。本研究评估了ChatGPT(一种会话人工智能模型)在提供哮喘相关信息方面的准确性和全面性。采用严格的混合方法,医疗保健专业人员评估了ChatGPT对成人哮喘常识问卷(AGKQA)的反应,AGKQA是一种涵盖各种哮喘相关主题的标准化工具。对回答的准确性和完整性进行分级,并使用统计测试进行分析,以评估再现性和一致性。ChatGPT在传达哮喘知识方面表现出显著的熟练程度,在病因和病理生理类别方面取得了完美的成功,在药物信息方面的准确性也很高(70%)。然而,在与药物相关的反应中注意到局限性,其中混合准确性(30%)强调需要进一步改进ChatGPT的能力,以确保哮喘教育关键领域的可靠性。可重复性分析表明,所有类别的可重复性都达到100%,这证实了ChatGPT在传递统一信息方面的可靠性。统计分析进一步强调了ChatGPT的稳定性和可靠性。这些发现强调了ChatGPT作为哮喘有价值的教育工具的前景,同时强调了持续改进的必要性,以解决观察到的局限性,特别是在药物相关信息方面。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Journal of Asthma
Journal of Asthma 医学-过敏
CiteScore
4.00
自引率
5.30%
发文量
158
审稿时长
3-8 weeks
期刊介绍: Providing an authoritative open forum on asthma and related conditions, Journal of Asthma publishes clinical research around such topics as asthma management, critical and long-term care, preventative measures, environmental counselling, and patient education.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信