Addressing the reporting chasm of artificial intelligence research: the DECIDE-AI reporting guidelines.

IF 2.1 Q2 SURGERY
BMJ Surgery Interventions Health Technologies Pub Date : 2022-07-29 eCollection Date: 2022-01-01 DOI:10.1136/bmjsit-2022-000154
John Gerrard Hanrahan, Danyal Zaman Khan, Hani J Marcus
{"title":"Addressing the reporting chasm of artificial intelligence research: the DECIDE-AI reporting guidelines.","authors":"John Gerrard Hanrahan, Danyal Zaman Khan, Hani J Marcus","doi":"10.1136/bmjsit-2022-000154","DOIUrl":null,"url":null,"abstract":"© Author(s) (or their employer(s)) 2022. Reuse permitted under CC BYNC. No commercial reuse. See rights and permissions. Published by BMJ. EDITORIAL The meteoric rise of artificial intelligence (AI) to the forefront of healthcare innovation has unearthed an array of avenues for surgical researchers to pursue. Applications found throughout the surgical patient pathway mean AI offers newfound support systems for clinical decisionmaking. Indeed, a growing number of technologies are entering clinical practice, with a recent review evaluating randomised controlled trials of diagnostic prediction tools suggests that potential benefits of AI that contemporary healthcare stands to realise. However, the pathway to translation to the bedside for these technologies is variable. Captured aptly in a recent editorial, there are clear examples of AI technologies already approved for clinical use in the USA, both with and without evaluation through randomised controlled trials. This speaks to a wider problem of evaluation in AI innovation, where insufficient reporting in randomised controlled trials prompted the development of several reporting guidelines, examples including the Consolidated Standards of Reporting TrialsAI and Standard Protocol Items: Recommendations for Interventional TrialsAI guidelines advising the minimum reporting standards for clinical trials and protocols, respectively. Similarly, guidance for the initial stages of AI development has been developed, namely, the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPODAI) guidelines for machine learning (ML) prediction models. Yet, when one looks at the process of AI translation, from in silico to clinical trial, an evaluation chasm becomes obvious, with guidance lacking on studies reflecting stages 2a and 2b of the IDEAL (Idea, Development, Exploration, Assessment, Longterm study) collaborative. These stages reflect the refinement and preparation for larger clinical studies, which are influenced by factors from the operator including learning curves or training; the health system the technologies enter into or organisational factors such as integration into clinical workflows. Study design features such as patient selection for both training and testing an intervention, and even the AI model itself, are crucial factors to consider prior to largescale testing. Vasey and colleagues have identified a gap in the reporting guidelines for evaluating AIdriven decision support systems, producing reporting guidelines to support the evaluation of their early stages. This was achieved through an international, tworound modified Delphi consensus process producing a 17 AIspecific item and 10 generic item reporting guidelines (DECIDEAI), informing the reporting of earlystage clinical studies of AIbased decision support systems in healthcare. The systems perspective taken by Vasey et al frame AI decisionsupport systems as complex interventions. This perspective clearly elucidates the importance of understanding of the workflow or clinical process interventions are intended to enter, alongside the evaluation setting of the AI. Reporting of such demonstrates the setting, or even systemspecific evaluation in the selected trial which may be important in judging intervention efficacy when applied to the same clinical problem in alternate health systems or settings. Furthermore, the emulation of aviation or military human factors appraisal is another value of the DECIDEAI guidelines, particularly as the augmentative nature of AI decisionsupport systems rely on humancomputer interactions. It is evident, for example, in surgery that learningcurves of surgeons influence clinical outcomes, meaning complex interventions including AIbased tools must account for this during coright.","PeriodicalId":33349,"journal":{"name":"BMJ Surgery Interventions Health Technologies","volume":null,"pages":null},"PeriodicalIF":2.1000,"publicationDate":"2022-07-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://ftp.ncbi.nlm.nih.gov/pub/pmc/oa_pdf/51/52/bmjsit-2022-000154.PMC9345081.pdf","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"BMJ Surgery Interventions Health Technologies","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1136/bmjsit-2022-000154","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2022/1/1 0:00:00","PubModel":"eCollection","JCR":"Q2","JCRName":"SURGERY","Score":null,"Total":0}
引用次数: 0

Abstract

© Author(s) (or their employer(s)) 2022. Reuse permitted under CC BYNC. No commercial reuse. See rights and permissions. Published by BMJ. EDITORIAL The meteoric rise of artificial intelligence (AI) to the forefront of healthcare innovation has unearthed an array of avenues for surgical researchers to pursue. Applications found throughout the surgical patient pathway mean AI offers newfound support systems for clinical decisionmaking. Indeed, a growing number of technologies are entering clinical practice, with a recent review evaluating randomised controlled trials of diagnostic prediction tools suggests that potential benefits of AI that contemporary healthcare stands to realise. However, the pathway to translation to the bedside for these technologies is variable. Captured aptly in a recent editorial, there are clear examples of AI technologies already approved for clinical use in the USA, both with and without evaluation through randomised controlled trials. This speaks to a wider problem of evaluation in AI innovation, where insufficient reporting in randomised controlled trials prompted the development of several reporting guidelines, examples including the Consolidated Standards of Reporting TrialsAI and Standard Protocol Items: Recommendations for Interventional TrialsAI guidelines advising the minimum reporting standards for clinical trials and protocols, respectively. Similarly, guidance for the initial stages of AI development has been developed, namely, the Transparent Reporting of a multivariable prediction model for Individual Prognosis Or Diagnosis (TRIPODAI) guidelines for machine learning (ML) prediction models. Yet, when one looks at the process of AI translation, from in silico to clinical trial, an evaluation chasm becomes obvious, with guidance lacking on studies reflecting stages 2a and 2b of the IDEAL (Idea, Development, Exploration, Assessment, Longterm study) collaborative. These stages reflect the refinement and preparation for larger clinical studies, which are influenced by factors from the operator including learning curves or training; the health system the technologies enter into or organisational factors such as integration into clinical workflows. Study design features such as patient selection for both training and testing an intervention, and even the AI model itself, are crucial factors to consider prior to largescale testing. Vasey and colleagues have identified a gap in the reporting guidelines for evaluating AIdriven decision support systems, producing reporting guidelines to support the evaluation of their early stages. This was achieved through an international, tworound modified Delphi consensus process producing a 17 AIspecific item and 10 generic item reporting guidelines (DECIDEAI), informing the reporting of earlystage clinical studies of AIbased decision support systems in healthcare. The systems perspective taken by Vasey et al frame AI decisionsupport systems as complex interventions. This perspective clearly elucidates the importance of understanding of the workflow or clinical process interventions are intended to enter, alongside the evaluation setting of the AI. Reporting of such demonstrates the setting, or even systemspecific evaluation in the selected trial which may be important in judging intervention efficacy when applied to the same clinical problem in alternate health systems or settings. Furthermore, the emulation of aviation or military human factors appraisal is another value of the DECIDEAI guidelines, particularly as the augmentative nature of AI decisionsupport systems rely on humancomputer interactions. It is evident, for example, in surgery that learningcurves of surgeons influence clinical outcomes, meaning complex interventions including AIbased tools must account for this during coright.
解决人工智能研究报告的鸿沟:decision - ai报告指南。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
CiteScore
2.80
自引率
0.00%
发文量
22
审稿时长
17 weeks
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信