chatgpt在回答患者预防性心脏病学问题中的质量和可读性:一项初步研究

IF 5.9 Q1 CARDIAC & CARDIOVASCULAR SYSTEMS
Prem Patel MD , Allison Bigeh DO , Benjamin Romer MD , Shantanu Dev BS , Samar Binkheder PhD , Lang Li PhD , Weidan Cao PhD , M. Wesley Milks MD
{"title":"chatgpt在回答患者预防性心脏病学问题中的质量和可读性:一项初步研究","authors":"Prem Patel MD ,&nbsp;Allison Bigeh DO ,&nbsp;Benjamin Romer MD ,&nbsp;Shantanu Dev BS ,&nbsp;Samar Binkheder PhD ,&nbsp;Lang Li PhD ,&nbsp;Weidan Cao PhD ,&nbsp;M. Wesley Milks MD","doi":"10.1016/j.ajpc.2025.101170","DOIUrl":null,"url":null,"abstract":"<div><h3>Therapeutic Area</h3><div>Other</div></div><div><h3>Background</h3><div>As artificial intelligence (AI) becomes increasingly integrated into healthcare, ChatGPT has emerged as a promising tool for patient education. However, research on its suitability for preventive cardiology remains limited. With patients increasingly relying on online health information, it is essential that content is both scientifically accurate and accessible to individuals of all health literacy levels. This study evaluates the quality and readability of ChatGPT’s responses to common questions on lifestyle modification, women’s cardiovascular health, and cholesterol management.</div></div><div><h3>Methods</h3><div>Twenty-six questions (8 on lifestyle modifications, 8 on women’s cardiovascular health, and 10 on cholesterol management) were queried using the GPT-4 model. Responses were independently evaluated by three board-certified preventive cardiologists, referencing the latest national cardiovascular guidelines. Quality was assessed using a 5-point Likert scale for correctness, comprehensiveness, conciseness, and comprehensibility, previously employed in medical AI research. Readability was analyzed using the Flesch-Kincaid Grade Level and other standardized readability metrics.</div></div><div><h3>Results</h3><div>ChatGPT provided adequate responses to 88.4% (23/26) of questions, with mean (SE) scores of 3.71 ± 0.20 for correctness, 4.06 ± 0.14 for conciseness, 4.06 ± 0.13 for comprehensiveness, and 4.40 ± 0.10 for comprehensibility. The highest-scoring topic was lifestyle modification (84.4%), followed by cholesterol management (81.2%) and women’s cardiovascular health (77.8%). Among inadequate responses, key limitations included overstating the risks of low LDL cholesterol and exaggerating the benefits of estrogen replacement therapy (ERT) for postmenopausal CVD risk reduction. ChatGPT also provided unsupported recommendations regarding dietary supplements for CVD prevention. Readability analysis revealed responses at a 13th-grade level, exceeding the recommended 6th-grade level for patient education.</div></div><div><h3>Conclusions</h3><div>ChatGPT’s responses were generally suitable for topics such as heart-healthy diets, exercise, weight management, epidemiology and clinical presentation of CVD in women, postmenopausal CVD risk, cholesterol-lowering therapy, statin-associated side effects, and Lp(a) risk stratification. However, inaccuracies persisted in dietary supplementation, ERT, and very low LDL levels. Enhancements in AI training are needed to improve accuracy in these areas. Additionally, the high readability level limits accessibility for the general public, underscoring the need for optimization to ensure clear and reliable patient education.</div></div>","PeriodicalId":72173,"journal":{"name":"American journal of preventive cardiology","volume":"23 ","pages":"Article 101170"},"PeriodicalIF":5.9000,"publicationDate":"2025-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"QUALITY AND READABILITY OF CHATGPT IN ANSWERING PATIENTS’ PREVENTIVE CARDIOLOGY QUESTIONS: A PILOT STUDY\",\"authors\":\"Prem Patel MD ,&nbsp;Allison Bigeh DO ,&nbsp;Benjamin Romer MD ,&nbsp;Shantanu Dev BS ,&nbsp;Samar Binkheder PhD ,&nbsp;Lang Li PhD ,&nbsp;Weidan Cao PhD ,&nbsp;M. Wesley Milks MD\",\"doi\":\"10.1016/j.ajpc.2025.101170\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><h3>Therapeutic Area</h3><div>Other</div></div><div><h3>Background</h3><div>As artificial intelligence (AI) becomes increasingly integrated into healthcare, ChatGPT has emerged as a promising tool for patient education. However, research on its suitability for preventive cardiology remains limited. With patients increasingly relying on online health information, it is essential that content is both scientifically accurate and accessible to individuals of all health literacy levels. This study evaluates the quality and readability of ChatGPT’s responses to common questions on lifestyle modification, women’s cardiovascular health, and cholesterol management.</div></div><div><h3>Methods</h3><div>Twenty-six questions (8 on lifestyle modifications, 8 on women’s cardiovascular health, and 10 on cholesterol management) were queried using the GPT-4 model. Responses were independently evaluated by three board-certified preventive cardiologists, referencing the latest national cardiovascular guidelines. Quality was assessed using a 5-point Likert scale for correctness, comprehensiveness, conciseness, and comprehensibility, previously employed in medical AI research. Readability was analyzed using the Flesch-Kincaid Grade Level and other standardized readability metrics.</div></div><div><h3>Results</h3><div>ChatGPT provided adequate responses to 88.4% (23/26) of questions, with mean (SE) scores of 3.71 ± 0.20 for correctness, 4.06 ± 0.14 for conciseness, 4.06 ± 0.13 for comprehensiveness, and 4.40 ± 0.10 for comprehensibility. The highest-scoring topic was lifestyle modification (84.4%), followed by cholesterol management (81.2%) and women’s cardiovascular health (77.8%). Among inadequate responses, key limitations included overstating the risks of low LDL cholesterol and exaggerating the benefits of estrogen replacement therapy (ERT) for postmenopausal CVD risk reduction. ChatGPT also provided unsupported recommendations regarding dietary supplements for CVD prevention. Readability analysis revealed responses at a 13th-grade level, exceeding the recommended 6th-grade level for patient education.</div></div><div><h3>Conclusions</h3><div>ChatGPT’s responses were generally suitable for topics such as heart-healthy diets, exercise, weight management, epidemiology and clinical presentation of CVD in women, postmenopausal CVD risk, cholesterol-lowering therapy, statin-associated side effects, and Lp(a) risk stratification. However, inaccuracies persisted in dietary supplementation, ERT, and very low LDL levels. Enhancements in AI training are needed to improve accuracy in these areas. Additionally, the high readability level limits accessibility for the general public, underscoring the need for optimization to ensure clear and reliable patient education.</div></div>\",\"PeriodicalId\":72173,\"journal\":{\"name\":\"American journal of preventive cardiology\",\"volume\":\"23 \",\"pages\":\"Article 101170\"},\"PeriodicalIF\":5.9000,\"publicationDate\":\"2025-09-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"American journal of preventive cardiology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2666667725002454\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"CARDIAC & CARDIOVASCULAR SYSTEMS\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"American journal of preventive cardiology","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2666667725002454","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"CARDIAC & CARDIOVASCULAR SYSTEMS","Score":null,"Total":0}
引用次数: 0

摘要

随着人工智能(AI)越来越多地融入医疗保健领域,ChatGPT已成为一种有前途的患者教育工具。然而,对其在预防心脏病学中的适用性的研究仍然有限。随着患者越来越依赖在线健康信息,内容必须在科学上准确,并且所有卫生知识水平的个人都可以访问。本研究评估了ChatGPT对生活方式改变、女性心血管健康和胆固醇管理等常见问题的回答的质量和可读性。方法采用GPT-4模型对26个问题(生活方式改变8个、女性心血管健康8个、胆固醇管理10个)进行问卷调查。三个委员会认证的预防心脏病专家参照最新的国家心血管指南,独立评估了这些反应。质量评估使用5分李克特量表,用于正确性、全面性、简洁性和可理解性,以前用于医疗人工智能研究。使用Flesch-Kincaid Grade Level和其他标准化的可读性指标分析可读性。结果atgpt对88.4%(23/26)的问题给出了充分的回答,正确性(正确性)为3.71±0.20分,简洁性(简洁性)为4.06±0.14分,全面性(全面性)为4.06±0.13分,可理解性(可理解性)为4.40±0.10分。得分最高的话题是生活方式的改变(84.4%),其次是胆固醇管理(81.2%)和女性心血管健康(77.8%)。在不充分的反应中,主要的局限性包括夸大低LDL胆固醇的风险和夸大雌激素替代疗法(ERT)降低绝经后心血管疾病风险的益处。ChatGPT还提供了关于预防心血管疾病的膳食补充剂的不受支持的建议。可读性分析显示,患者的反应水平为13年级,超过了患者教育建议的6年级水平。结论:schatgpt的应答通常适用于心脏健康饮食、运动、体重管理、女性心血管疾病的流行病学和临床表现、绝经后心血管疾病风险、降胆固醇治疗、他汀类药物相关副作用和Lp(a)风险分层等主题。然而,在膳食补充、ERT和非常低的LDL水平中,不准确性仍然存在。需要加强人工智能训练,以提高这些领域的准确性。此外,高可读性水平限制了普通公众的可访问性,强调了优化的必要性,以确保清晰可靠的患者教育。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
QUALITY AND READABILITY OF CHATGPT IN ANSWERING PATIENTS’ PREVENTIVE CARDIOLOGY QUESTIONS: A PILOT STUDY

Therapeutic Area

Other

Background

As artificial intelligence (AI) becomes increasingly integrated into healthcare, ChatGPT has emerged as a promising tool for patient education. However, research on its suitability for preventive cardiology remains limited. With patients increasingly relying on online health information, it is essential that content is both scientifically accurate and accessible to individuals of all health literacy levels. This study evaluates the quality and readability of ChatGPT’s responses to common questions on lifestyle modification, women’s cardiovascular health, and cholesterol management.

Methods

Twenty-six questions (8 on lifestyle modifications, 8 on women’s cardiovascular health, and 10 on cholesterol management) were queried using the GPT-4 model. Responses were independently evaluated by three board-certified preventive cardiologists, referencing the latest national cardiovascular guidelines. Quality was assessed using a 5-point Likert scale for correctness, comprehensiveness, conciseness, and comprehensibility, previously employed in medical AI research. Readability was analyzed using the Flesch-Kincaid Grade Level and other standardized readability metrics.

Results

ChatGPT provided adequate responses to 88.4% (23/26) of questions, with mean (SE) scores of 3.71 ± 0.20 for correctness, 4.06 ± 0.14 for conciseness, 4.06 ± 0.13 for comprehensiveness, and 4.40 ± 0.10 for comprehensibility. The highest-scoring topic was lifestyle modification (84.4%), followed by cholesterol management (81.2%) and women’s cardiovascular health (77.8%). Among inadequate responses, key limitations included overstating the risks of low LDL cholesterol and exaggerating the benefits of estrogen replacement therapy (ERT) for postmenopausal CVD risk reduction. ChatGPT also provided unsupported recommendations regarding dietary supplements for CVD prevention. Readability analysis revealed responses at a 13th-grade level, exceeding the recommended 6th-grade level for patient education.

Conclusions

ChatGPT’s responses were generally suitable for topics such as heart-healthy diets, exercise, weight management, epidemiology and clinical presentation of CVD in women, postmenopausal CVD risk, cholesterol-lowering therapy, statin-associated side effects, and Lp(a) risk stratification. However, inaccuracies persisted in dietary supplementation, ERT, and very low LDL levels. Enhancements in AI training are needed to improve accuracy in these areas. Additionally, the high readability level limits accessibility for the general public, underscoring the need for optimization to ensure clear and reliable patient education.
求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
American journal of preventive cardiology
American journal of preventive cardiology Cardiology and Cardiovascular Medicine
CiteScore
6.60
自引率
0.00%
发文量
0
审稿时长
76 days
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信