Quality and Readability Analysis of Artificial Intelligence-Generated Medical Information Related to Prostate Cancer: A Cross-Sectional Study of ChatGPT and DeepSeek.

IF 4.1 3区 医学 Q1 ANDROLOGY
Zhao Luo, Chuan Lin, Tae Hyo Kim, Yu Seob Shin, Sun Tae Ahn
{"title":"Quality and Readability Analysis of Artificial Intelligence-Generated Medical Information Related to Prostate Cancer: A Cross-Sectional Study of ChatGPT and DeepSeek.","authors":"Zhao Luo, Chuan Lin, Tae Hyo Kim, Yu Seob Shin, Sun Tae Ahn","doi":"10.5534/wjmh.250144","DOIUrl":null,"url":null,"abstract":"<p><strong>Purpose: </strong>Artificial intelligence (AI) tools have demonstrated considerable potential for the dissemination of medical information. However, variability may exist in the quality and readability of prostate-cancer-related content generated by different AI platforms. This study aimed to evaluate the quality, accuracy, and readability of prostate-cancer-related medical information produced by ChatGPT and DeepSeek.</p><p><strong>Materials and methods: </strong>Frequently asked questions related to prostate cancer were collected from the American Cancer Society website, ChatGPT, and DeepSeek. Three urologists with over 10 years of clinical experience reviewed and confirmed the relevance of the selected questions. The Patient Education Materials Assessment Tool for Printable Materials (PEMAT-P) was used to assess the understandability and actionability of AI-generated content. The DISCERN instrument was used to evaluate the quality of the treatment-related information. Additionally, readability was assessed using four established indices: Automated Readability Index (ARI), Flesch Reading Ease Score, Gunning Fog Index, and Flesch-Kincaid Grade Level.</p><p><strong>Results: </strong>No statistically significant differences were observed between ChatGPT and DeepSeek in PEMAT-P scores (70.66±8.13 <i>vs</i>. 69.35±8.83) or DISCERN scores (59.07±3.39 <i>vs</i>. 58.88±3.66) (p>0.05). However, the ARI for DeepSeek was higher than that for ChatGPT (12.63±1.42 <i>vs</i>. 10.85±1.93, p<0.001), indicating greater textual complexity and reading difficulty.</p><p><strong>Conclusions: </strong>AI tools, such as ChatGPT and DeepSeek, hold significant potential for enhancing patient education and disseminating medical information on prostate cancer. Nevertheless, further refinement of content quality and language clarity is needed to prevent potential misunderstandings, decisional uncertainty, and anxiety among patients due to difficulty in comprehension.</p>","PeriodicalId":54261,"journal":{"name":"World Journal of Mens Health","volume":" ","pages":""},"PeriodicalIF":4.1000,"publicationDate":"2025-09-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"World Journal of Mens Health","FirstCategoryId":"3","ListUrlMain":"https://doi.org/10.5534/wjmh.250144","RegionNum":3,"RegionCategory":"医学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ANDROLOGY","Score":null,"Total":0}
引用次数: 0

Abstract

Purpose: Artificial intelligence (AI) tools have demonstrated considerable potential for the dissemination of medical information. However, variability may exist in the quality and readability of prostate-cancer-related content generated by different AI platforms. This study aimed to evaluate the quality, accuracy, and readability of prostate-cancer-related medical information produced by ChatGPT and DeepSeek.

Materials and methods: Frequently asked questions related to prostate cancer were collected from the American Cancer Society website, ChatGPT, and DeepSeek. Three urologists with over 10 years of clinical experience reviewed and confirmed the relevance of the selected questions. The Patient Education Materials Assessment Tool for Printable Materials (PEMAT-P) was used to assess the understandability and actionability of AI-generated content. The DISCERN instrument was used to evaluate the quality of the treatment-related information. Additionally, readability was assessed using four established indices: Automated Readability Index (ARI), Flesch Reading Ease Score, Gunning Fog Index, and Flesch-Kincaid Grade Level.

Results: No statistically significant differences were observed between ChatGPT and DeepSeek in PEMAT-P scores (70.66±8.13 vs. 69.35±8.83) or DISCERN scores (59.07±3.39 vs. 58.88±3.66) (p>0.05). However, the ARI for DeepSeek was higher than that for ChatGPT (12.63±1.42 vs. 10.85±1.93, p<0.001), indicating greater textual complexity and reading difficulty.

Conclusions: AI tools, such as ChatGPT and DeepSeek, hold significant potential for enhancing patient education and disseminating medical information on prostate cancer. Nevertheless, further refinement of content quality and language clarity is needed to prevent potential misunderstandings, decisional uncertainty, and anxiety among patients due to difficulty in comprehension.

人工智能生成的前列腺癌相关医疗信息的质量和可读性分析:ChatGPT和DeepSeek的横断面研究
目的:人工智能(AI)工具已显示出传播医疗信息的巨大潜力。然而,不同人工智能平台生成的前列腺癌相关内容的质量和可读性可能存在差异。本研究旨在评估ChatGPT和DeepSeek提供的前列腺癌相关医学信息的质量、准确性和可读性。材料和方法:从美国癌症协会网站、ChatGPT和DeepSeek收集与前列腺癌相关的常见问题。三位具有超过10年临床经验的泌尿科医生审查并确认了所选问题的相关性。使用可打印材料患者教育材料评估工具(PEMAT-P)来评估人工智能生成内容的可理解性和可操作性。使用DISCERN仪器来评估治疗相关信息的质量。此外,使用4个既定指标评估可读性:自动可读性指数(ARI)、Flesch阅读简易评分、Gunning Fog指数和Flesch- kincaid Grade Level。结果:ChatGPT和DeepSeek在PEMAT-P评分(70.66±8.13比69.35±8.83)和DISCERN评分(59.07±3.39比58.88±3.66)方面差异均无统计学意义(p < 0.05)。然而,DeepSeek的ARI高于ChatGPT(12.63±1.42 vs. 10.85±1.93)。结论:人工智能工具,如ChatGPT和DeepSeek,在加强患者教育和传播前列腺癌医学信息方面具有重要潜力。然而,需要进一步改进内容质量和语言清晰度,以防止患者因理解困难而产生误解、决策不确定性和焦虑。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
World Journal of Mens Health
World Journal of Mens Health Medicine-Psychiatry and Mental Health
CiteScore
7.60
自引率
2.10%
发文量
92
审稿时长
6 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信