MedReadr:浏览器内基于规则的自然语言处理算法的开发和评估,用于估计消费者健康文章的可靠性。

Biomedicine hub Pub Date : 2025-08-25 eCollection Date: 2025-01-01 DOI:10.1159/000548163
Joshua Winograd, Autumn Kim, Nikit Venishetty, Alia Codelia-Anjum, Dean Elterman, Naeem Bhojani, Kevin C Zorn, Adithya Balasubramanian, Andrew Vickers, Bilal Chughtai
{"title":"MedReadr:浏览器内基于规则的自然语言处理算法的开发和评估,用于估计消费者健康文章的可靠性。","authors":"Joshua Winograd, Autumn Kim, Nikit Venishetty, Alia Codelia-Anjum, Dean Elterman, Naeem Bhojani, Kevin C Zorn, Adithya Balasubramanian, Andrew Vickers, Bilal Chughtai","doi":"10.1159/000548163","DOIUrl":null,"url":null,"abstract":"<p><strong>Introduction: </strong>The internet is a major source of medical information for patients, yet the quality of online health content remains highly variable. Existing assessment tools are often labor-intensive, invalidated, or limited in scope. We developed and validated MedReadr, an in-browser, rule-based natural language processing (NLP) algorithm that automatically estimates the reliability of consumer health articles for patients and providers.</p><p><strong>Methods: </strong>Thirty-five consumer medical articles were independently assessed by two reviewers using validated manual scoring systems (QUEST and Sandvik). Interrater reliability was evaluated with Cohen's κ, and metrics with κ > 0.6 were selected for model fitting. MedReadr extracted key features from article text and metadata using predefined NLP rules. A multivariable linear regression model was trained to predict manual reliability scores, with internal validation performed on an independent set of 20 articles.</p><p><strong>Results: </strong>High interrater reliability was achieved across all QUEST and most Sandvik domains (Cohen's κ > 0.6). The MedReadr model demonstrated strong performance, achieving <i>R</i> <sup>2</sup> = 0.90 and RMSE = 0.05 on the development set and <i>R</i> <sup>2</sup> = 0.83 and RMSE = 0.07 on the validation set. All model coefficients were statistically significant (<i>p</i> < 0.05). Key predictive features included currency and reference scores, sentiment polarity, engagement content, and the frequency of provider contact, intervention endorsement, intervention mechanism, and intervention uncertainty phrases.</p><p><strong>Conclusion: </strong>MedReadr demonstrates that structural reliability scoring of online health articles can be automated using a transparent, rule-based NLP approach. Applied to English-language articles from mainstream search results on common medical conditions, the tool showed strong agreement with validated manual scoring systems. However, it has only been validated on a narrow scope of content and is not designed to analyze search results for specific questions or detect misinformation. Future research should assess its performance across a broader range of web content and evaluate whether its integration improves patient comprehension, digital health literacy, and clinician-patient communication.</p>","PeriodicalId":101351,"journal":{"name":"Biomedicine hub","volume":"10 1","pages":"162-170"},"PeriodicalIF":0.0000,"publicationDate":"2025-08-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12503773/pdf/","citationCount":"0","resultStr":"{\"title\":\"MedReadr: Development and Evaluation of an In-Browser, Rule-Based Natural Language Processing Algorithm to Estimate the Reliability of Consumer Health Articles.\",\"authors\":\"Joshua Winograd, Autumn Kim, Nikit Venishetty, Alia Codelia-Anjum, Dean Elterman, Naeem Bhojani, Kevin C Zorn, Adithya Balasubramanian, Andrew Vickers, Bilal Chughtai\",\"doi\":\"10.1159/000548163\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Introduction: </strong>The internet is a major source of medical information for patients, yet the quality of online health content remains highly variable. Existing assessment tools are often labor-intensive, invalidated, or limited in scope. We developed and validated MedReadr, an in-browser, rule-based natural language processing (NLP) algorithm that automatically estimates the reliability of consumer health articles for patients and providers.</p><p><strong>Methods: </strong>Thirty-five consumer medical articles were independently assessed by two reviewers using validated manual scoring systems (QUEST and Sandvik). Interrater reliability was evaluated with Cohen's κ, and metrics with κ > 0.6 were selected for model fitting. MedReadr extracted key features from article text and metadata using predefined NLP rules. A multivariable linear regression model was trained to predict manual reliability scores, with internal validation performed on an independent set of 20 articles.</p><p><strong>Results: </strong>High interrater reliability was achieved across all QUEST and most Sandvik domains (Cohen's κ > 0.6). The MedReadr model demonstrated strong performance, achieving <i>R</i> <sup>2</sup> = 0.90 and RMSE = 0.05 on the development set and <i>R</i> <sup>2</sup> = 0.83 and RMSE = 0.07 on the validation set. All model coefficients were statistically significant (<i>p</i> < 0.05). Key predictive features included currency and reference scores, sentiment polarity, engagement content, and the frequency of provider contact, intervention endorsement, intervention mechanism, and intervention uncertainty phrases.</p><p><strong>Conclusion: </strong>MedReadr demonstrates that structural reliability scoring of online health articles can be automated using a transparent, rule-based NLP approach. Applied to English-language articles from mainstream search results on common medical conditions, the tool showed strong agreement with validated manual scoring systems. However, it has only been validated on a narrow scope of content and is not designed to analyze search results for specific questions or detect misinformation. Future research should assess its performance across a broader range of web content and evaluate whether its integration improves patient comprehension, digital health literacy, and clinician-patient communication.</p>\",\"PeriodicalId\":101351,\"journal\":{\"name\":\"Biomedicine hub\",\"volume\":\"10 1\",\"pages\":\"162-170\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2025-08-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12503773/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Biomedicine hub\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1159/000548163\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/1/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Biomedicine hub","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1159/000548163","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/1/1 0:00:00","PubModel":"eCollection","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

导读:互联网是患者医疗信息的主要来源,但在线健康内容的质量仍然高度可变。现有的评估工具通常是劳动密集型的,无效的,或者范围有限的。我们开发并验证了MedReadr,这是一种基于规则的浏览器内自然语言处理(NLP)算法,可以自动估计患者和提供者的消费者健康文章的可靠性。方法:35篇消费类医学文章由两位审稿人使用经过验证的人工评分系统(QUEST和Sandvik)独立评估。采用Cohen’s κ进行信度评估,选择κ > 0.6的指标进行模型拟合。MedReadr使用预定义的NLP规则从文章文本和元数据中提取关键特征。训练多变量线性回归模型来预测人工信度评分,并对20篇文章的独立集进行内部验证。结果:在所有QUEST和大多数Sandvik域(Cohen’s κ > 0.6)中实现了高互译信度。MedReadr模型表现出较强的性能,在开发集上实现r2 = 0.90, RMSE = 0.05,在验证集上实现r2 = 0.83, RMSE = 0.07。各模型系数均有统计学意义(p < 0.05)。主要预测特征包括货币和参考分数、情绪极性、参与内容、提供者联系频率、干预认可、干预机制和干预不确定性短语。结论:medreader表明,在线健康文章的结构可靠性评分可以使用透明的、基于规则的NLP方法自动进行。应用于主流搜索结果中关于常见医疗条件的英文文章,该工具与经过验证的手动评分系统表现出强烈的一致性。然而,它只在一个狭窄的内容范围内进行了验证,并不是为了分析特定问题的搜索结果或检测错误信息而设计的。未来的研究应评估其在更广泛的网络内容上的表现,并评估其整合是否能改善患者理解、数字健康素养和医患沟通。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

MedReadr: Development and Evaluation of an In-Browser, Rule-Based Natural Language Processing Algorithm to Estimate the Reliability of Consumer Health Articles.

MedReadr: Development and Evaluation of an In-Browser, Rule-Based Natural Language Processing Algorithm to Estimate the Reliability of Consumer Health Articles.

MedReadr: Development and Evaluation of an In-Browser, Rule-Based Natural Language Processing Algorithm to Estimate the Reliability of Consumer Health Articles.

Introduction: The internet is a major source of medical information for patients, yet the quality of online health content remains highly variable. Existing assessment tools are often labor-intensive, invalidated, or limited in scope. We developed and validated MedReadr, an in-browser, rule-based natural language processing (NLP) algorithm that automatically estimates the reliability of consumer health articles for patients and providers.

Methods: Thirty-five consumer medical articles were independently assessed by two reviewers using validated manual scoring systems (QUEST and Sandvik). Interrater reliability was evaluated with Cohen's κ, and metrics with κ > 0.6 were selected for model fitting. MedReadr extracted key features from article text and metadata using predefined NLP rules. A multivariable linear regression model was trained to predict manual reliability scores, with internal validation performed on an independent set of 20 articles.

Results: High interrater reliability was achieved across all QUEST and most Sandvik domains (Cohen's κ > 0.6). The MedReadr model demonstrated strong performance, achieving R 2 = 0.90 and RMSE = 0.05 on the development set and R 2 = 0.83 and RMSE = 0.07 on the validation set. All model coefficients were statistically significant (p < 0.05). Key predictive features included currency and reference scores, sentiment polarity, engagement content, and the frequency of provider contact, intervention endorsement, intervention mechanism, and intervention uncertainty phrases.

Conclusion: MedReadr demonstrates that structural reliability scoring of online health articles can be automated using a transparent, rule-based NLP approach. Applied to English-language articles from mainstream search results on common medical conditions, the tool showed strong agreement with validated manual scoring systems. However, it has only been validated on a narrow scope of content and is not designed to analyze search results for specific questions or detect misinformation. Future research should assess its performance across a broader range of web content and evaluate whether its integration improves patient comprehension, digital health literacy, and clinician-patient communication.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信