Assessing the Reporting Quality of Machine Learning Algorithms in Head and Neck Oncology

Rahul Alapati, Bryan Renslo, Sarah F. Wagoner, Omar Karadaghy, Aisha Serpedin, Yeo Eun Kim, Maria Feucht, Naomi Wang, Uma Ramesh, Antonio Bon Nieves, Amelia Lawrence, Celina Virgen, Tuleen Sawaf, Anaïs Rameau, Andrés M. Bur
{"title":"Assessing the Reporting Quality of Machine Learning Algorithms in Head and Neck Oncology","authors":"Rahul Alapati, Bryan Renslo, Sarah F. Wagoner, Omar Karadaghy, Aisha Serpedin, Yeo Eun Kim, Maria Feucht, Naomi Wang, Uma Ramesh, Antonio Bon Nieves, Amelia Lawrence, Celina Virgen, Tuleen Sawaf, Anaïs Rameau, Andrés M. Bur","doi":"10.1002/lary.31756","DOIUrl":null,"url":null,"abstract":"ObjectiveThis study aimed to assess reporting quality of machine learning (ML) algorithms in the head and neck oncology literature using the TRIPOD‐AI criteria.Data SourcesA comprehensive search was conducted using PubMed, Scopus, Embase, and Cochrane Database of Systematic Reviews, incorporating search terms related to “artificial intelligence,” “machine learning,” “deep learning,” “neural network,” and various head and neck neoplasms.Review MethodsTwo independent reviewers analyzed each published study for adherence to the 65‐point TRIPOD‐AI criteria. Items were classified as “Yes,” “No,” or “NA” for each publication. The proportion of studies satisfying each TRIPOD‐AI criterion was calculated. Additionally, the evidence level for each study was evaluated independently by two reviewers using the Oxford Centre for Evidence‐Based Medicine (OCEBM) Levels of Evidence. Discrepancies were reconciled through discussion until consensus was reached.ResultsThe study highlights the need for improvements in ML algorithm reporting in head and neck oncology. This includes more comprehensive descriptions of datasets, standardization of model performance reporting, and increased sharing of ML models, data, and code with the research community. Adoption of TRIPOD‐AI is necessary for achieving standardized ML research reporting in head and neck oncology.ConclusionCurrent reporting of ML algorithms hinders clinical application, reproducibility, and understanding of the data used for model training. To overcome these limitations and improve patient and clinician trust, ML developers should provide open access to models, code, and source data, fostering iterative progress through community critique, thus enhancing model accuracy and mitigating biases.Level of EvidenceNA <jats:italic>Laryngoscope</jats:italic>, 2024","PeriodicalId":501696,"journal":{"name":"The Laryngoscope","volume":"24 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Laryngoscope","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1002/lary.31756","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

ObjectiveThis study aimed to assess reporting quality of machine learning (ML) algorithms in the head and neck oncology literature using the TRIPOD‐AI criteria.Data SourcesA comprehensive search was conducted using PubMed, Scopus, Embase, and Cochrane Database of Systematic Reviews, incorporating search terms related to “artificial intelligence,” “machine learning,” “deep learning,” “neural network,” and various head and neck neoplasms.Review MethodsTwo independent reviewers analyzed each published study for adherence to the 65‐point TRIPOD‐AI criteria. Items were classified as “Yes,” “No,” or “NA” for each publication. The proportion of studies satisfying each TRIPOD‐AI criterion was calculated. Additionally, the evidence level for each study was evaluated independently by two reviewers using the Oxford Centre for Evidence‐Based Medicine (OCEBM) Levels of Evidence. Discrepancies were reconciled through discussion until consensus was reached.ResultsThe study highlights the need for improvements in ML algorithm reporting in head and neck oncology. This includes more comprehensive descriptions of datasets, standardization of model performance reporting, and increased sharing of ML models, data, and code with the research community. Adoption of TRIPOD‐AI is necessary for achieving standardized ML research reporting in head and neck oncology.ConclusionCurrent reporting of ML algorithms hinders clinical application, reproducibility, and understanding of the data used for model training. To overcome these limitations and improve patient and clinician trust, ML developers should provide open access to models, code, and source data, fostering iterative progress through community critique, thus enhancing model accuracy and mitigating biases.Level of EvidenceNA Laryngoscope, 2024
评估头颈部肿瘤学中机器学习算法的报告质量
目的 本研究旨在使用 TRIPOD-AI 标准评估头颈部肿瘤文献中机器学习(ML)算法的报告质量。数据来源使用PubMed、Scopus、Embase和Cochrane系统性综述数据库进行了全面检索,纳入了与 "人工智能"、"机器学习"、"深度学习"、"神经网络 "和各种头颈部肿瘤相关的检索词。 评审方法两位独立评审员分析了每篇已发表的研究是否符合65点TRIPOD-AI标准。每篇论文的项目被分为 "是"、"否 "或 "不适用"。计算符合 TRIPOD-AI 各项标准的研究比例。此外,每项研究的证据级别均由两名审稿人使用牛津循证医学中心(OCEBM)的证据级别进行独立评估。研究结果该研究强调了改进头颈部肿瘤学 ML 算法报告的必要性。这包括对数据集进行更全面的描述,对模型性能报告进行标准化,以及加强与研究界共享 ML 模型、数据和代码。采用 TRIPOD-AI 对于实现头颈部肿瘤学领域 ML 研究报告的标准化非常必要。为了克服这些局限性并提高患者和临床医生的信任度,ML 开发人员应提供对模型、代码和源数据的开放访问权限,通过社区评论促进迭代进步,从而提高模型的准确性并减少偏差。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信