Rahul Alapati, Bryan Renslo, Sarah F. Wagoner, Omar Karadaghy, Aisha Serpedin, Yeo Eun Kim, Maria Feucht, Naomi Wang, Uma Ramesh, Antonio Bon Nieves, Amelia Lawrence, Celina Virgen, Tuleen Sawaf, Anaïs Rameau, Andrés M. Bur
{"title":"评估头颈部肿瘤学中机器学习算法的报告质量","authors":"Rahul Alapati, Bryan Renslo, Sarah F. Wagoner, Omar Karadaghy, Aisha Serpedin, Yeo Eun Kim, Maria Feucht, Naomi Wang, Uma Ramesh, Antonio Bon Nieves, Amelia Lawrence, Celina Virgen, Tuleen Sawaf, Anaïs Rameau, Andrés M. Bur","doi":"10.1002/lary.31756","DOIUrl":null,"url":null,"abstract":"ObjectiveThis study aimed to assess reporting quality of machine learning (ML) algorithms in the head and neck oncology literature using the TRIPOD‐AI criteria.Data SourcesA comprehensive search was conducted using PubMed, Scopus, Embase, and Cochrane Database of Systematic Reviews, incorporating search terms related to “artificial intelligence,” “machine learning,” “deep learning,” “neural network,” and various head and neck neoplasms.Review MethodsTwo independent reviewers analyzed each published study for adherence to the 65‐point TRIPOD‐AI criteria. Items were classified as “Yes,” “No,” or “NA” for each publication. The proportion of studies satisfying each TRIPOD‐AI criterion was calculated. Additionally, the evidence level for each study was evaluated independently by two reviewers using the Oxford Centre for Evidence‐Based Medicine (OCEBM) Levels of Evidence. Discrepancies were reconciled through discussion until consensus was reached.ResultsThe study highlights the need for improvements in ML algorithm reporting in head and neck oncology. This includes more comprehensive descriptions of datasets, standardization of model performance reporting, and increased sharing of ML models, data, and code with the research community. Adoption of TRIPOD‐AI is necessary for achieving standardized ML research reporting in head and neck oncology.ConclusionCurrent reporting of ML algorithms hinders clinical application, reproducibility, and understanding of the data used for model training. To overcome these limitations and improve patient and clinician trust, ML developers should provide open access to models, code, and source data, fostering iterative progress through community critique, thus enhancing model accuracy and mitigating biases.Level of EvidenceNA <jats:italic>Laryngoscope</jats:italic>, 2024","PeriodicalId":501696,"journal":{"name":"The Laryngoscope","volume":"24 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Assessing the Reporting Quality of Machine Learning Algorithms in Head and Neck Oncology\",\"authors\":\"Rahul Alapati, Bryan Renslo, Sarah F. Wagoner, Omar Karadaghy, Aisha Serpedin, Yeo Eun Kim, Maria Feucht, Naomi Wang, Uma Ramesh, Antonio Bon Nieves, Amelia Lawrence, Celina Virgen, Tuleen Sawaf, Anaïs Rameau, Andrés M. Bur\",\"doi\":\"10.1002/lary.31756\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"ObjectiveThis study aimed to assess reporting quality of machine learning (ML) algorithms in the head and neck oncology literature using the TRIPOD‐AI criteria.Data SourcesA comprehensive search was conducted using PubMed, Scopus, Embase, and Cochrane Database of Systematic Reviews, incorporating search terms related to “artificial intelligence,” “machine learning,” “deep learning,” “neural network,” and various head and neck neoplasms.Review MethodsTwo independent reviewers analyzed each published study for adherence to the 65‐point TRIPOD‐AI criteria. Items were classified as “Yes,” “No,” or “NA” for each publication. The proportion of studies satisfying each TRIPOD‐AI criterion was calculated. Additionally, the evidence level for each study was evaluated independently by two reviewers using the Oxford Centre for Evidence‐Based Medicine (OCEBM) Levels of Evidence. Discrepancies were reconciled through discussion until consensus was reached.ResultsThe study highlights the need for improvements in ML algorithm reporting in head and neck oncology. This includes more comprehensive descriptions of datasets, standardization of model performance reporting, and increased sharing of ML models, data, and code with the research community. Adoption of TRIPOD‐AI is necessary for achieving standardized ML research reporting in head and neck oncology.ConclusionCurrent reporting of ML algorithms hinders clinical application, reproducibility, and understanding of the data used for model training. To overcome these limitations and improve patient and clinician trust, ML developers should provide open access to models, code, and source data, fostering iterative progress through community critique, thus enhancing model accuracy and mitigating biases.Level of EvidenceNA <jats:italic>Laryngoscope</jats:italic>, 2024\",\"PeriodicalId\":501696,\"journal\":{\"name\":\"The Laryngoscope\",\"volume\":\"24 1\",\"pages\":\"\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-09-11\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"The Laryngoscope\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1002/lary.31756\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"The Laryngoscope","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1002/lary.31756","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
目的 本研究旨在使用 TRIPOD-AI 标准评估头颈部肿瘤文献中机器学习(ML)算法的报告质量。数据来源使用PubMed、Scopus、Embase和Cochrane系统性综述数据库进行了全面检索,纳入了与 "人工智能"、"机器学习"、"深度学习"、"神经网络 "和各种头颈部肿瘤相关的检索词。 评审方法两位独立评审员分析了每篇已发表的研究是否符合65点TRIPOD-AI标准。每篇论文的项目被分为 "是"、"否 "或 "不适用"。计算符合 TRIPOD-AI 各项标准的研究比例。此外,每项研究的证据级别均由两名审稿人使用牛津循证医学中心(OCEBM)的证据级别进行独立评估。研究结果该研究强调了改进头颈部肿瘤学 ML 算法报告的必要性。这包括对数据集进行更全面的描述,对模型性能报告进行标准化,以及加强与研究界共享 ML 模型、数据和代码。采用 TRIPOD-AI 对于实现头颈部肿瘤学领域 ML 研究报告的标准化非常必要。为了克服这些局限性并提高患者和临床医生的信任度,ML 开发人员应提供对模型、代码和源数据的开放访问权限,通过社区评论促进迭代进步,从而提高模型的准确性并减少偏差。
Assessing the Reporting Quality of Machine Learning Algorithms in Head and Neck Oncology
ObjectiveThis study aimed to assess reporting quality of machine learning (ML) algorithms in the head and neck oncology literature using the TRIPOD‐AI criteria.Data SourcesA comprehensive search was conducted using PubMed, Scopus, Embase, and Cochrane Database of Systematic Reviews, incorporating search terms related to “artificial intelligence,” “machine learning,” “deep learning,” “neural network,” and various head and neck neoplasms.Review MethodsTwo independent reviewers analyzed each published study for adherence to the 65‐point TRIPOD‐AI criteria. Items were classified as “Yes,” “No,” or “NA” for each publication. The proportion of studies satisfying each TRIPOD‐AI criterion was calculated. Additionally, the evidence level for each study was evaluated independently by two reviewers using the Oxford Centre for Evidence‐Based Medicine (OCEBM) Levels of Evidence. Discrepancies were reconciled through discussion until consensus was reached.ResultsThe study highlights the need for improvements in ML algorithm reporting in head and neck oncology. This includes more comprehensive descriptions of datasets, standardization of model performance reporting, and increased sharing of ML models, data, and code with the research community. Adoption of TRIPOD‐AI is necessary for achieving standardized ML research reporting in head and neck oncology.ConclusionCurrent reporting of ML algorithms hinders clinical application, reproducibility, and understanding of the data used for model training. To overcome these limitations and improve patient and clinician trust, ML developers should provide open access to models, code, and source data, fostering iterative progress through community critique, thus enhancing model accuracy and mitigating biases.Level of EvidenceNA Laryngoscope, 2024