AI in the Loop:功能化折叠性能差异,以监控自动医学图像分割工作流程。

Frontiers in radiology Pub Date : 2023-09-15 eCollection Date: 2023-01-01 DOI:10.3389/fradi.2023.1223294
Harrison C Gottlich, Panagiotis Korfiatis, Adriana V Gregory, Timothy L Kline
{"title":"AI in the Loop:功能化折叠性能差异,以监控自动医学图像分割工作流程。","authors":"Harrison C Gottlich, Panagiotis Korfiatis, Adriana V Gregory, Timothy L Kline","doi":"10.3389/fradi.2023.1223294","DOIUrl":null,"url":null,"abstract":"<p><strong>Introduction: </strong>Methods that automatically flag poor performing predictions are drastically needed to safely implement machine learning workflows into clinical practice as well as to identify difficult cases during model training.</p><p><strong>Methods: </strong>Disagreement between the fivefold cross-validation sub-models was quantified using dice scores between folds and summarized as a surrogate for model confidence. The summarized Interfold Dices were compared with thresholds informed by human interobserver values to determine whether final ensemble model performance should be manually reviewed.</p><p><strong>Results: </strong>The method on all tasks efficiently flagged poor segmented images without consulting a reference standard. Using the median Interfold Dice for comparison, substantial dice score improvements after excluding flagged images was noted for the in-domain CT (0.85 ± 0.20 to 0.91 ± 0.08, 8/50 images flagged) and MR (0.76 ± 0.27 to 0.85 ± 0.09, 8/50 images flagged). Most impressively, there were dramatic dice score improvements in the simulated out-of-distribution task where the model was trained on a radical nephrectomy dataset with different contrast phases predicting a partial nephrectomy all cortico-medullary phase dataset (0.67 ± 0.36 to 0.89 ± 0.10, 122/300 images flagged).</p><p><strong>Discussion: </strong>Comparing interfold sub-model disagreement against human interobserver values is an effective and efficient way to assess automated predictions when a reference standard is not available. This functionality provides a necessary safeguard to patient care important to safely implement automated medical image segmentation workflows.</p>","PeriodicalId":73101,"journal":{"name":"Frontiers in radiology","volume":"3 ","pages":"1223294"},"PeriodicalIF":0.0000,"publicationDate":"2023-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10540615/pdf/","citationCount":"0","resultStr":"{\"title\":\"AI in the Loop: functionalizing fold performance disagreement to monitor automated medical image segmentation workflows.\",\"authors\":\"Harrison C Gottlich, Panagiotis Korfiatis, Adriana V Gregory, Timothy L Kline\",\"doi\":\"10.3389/fradi.2023.1223294\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Introduction: </strong>Methods that automatically flag poor performing predictions are drastically needed to safely implement machine learning workflows into clinical practice as well as to identify difficult cases during model training.</p><p><strong>Methods: </strong>Disagreement between the fivefold cross-validation sub-models was quantified using dice scores between folds and summarized as a surrogate for model confidence. The summarized Interfold Dices were compared with thresholds informed by human interobserver values to determine whether final ensemble model performance should be manually reviewed.</p><p><strong>Results: </strong>The method on all tasks efficiently flagged poor segmented images without consulting a reference standard. Using the median Interfold Dice for comparison, substantial dice score improvements after excluding flagged images was noted for the in-domain CT (0.85 ± 0.20 to 0.91 ± 0.08, 8/50 images flagged) and MR (0.76 ± 0.27 to 0.85 ± 0.09, 8/50 images flagged). Most impressively, there were dramatic dice score improvements in the simulated out-of-distribution task where the model was trained on a radical nephrectomy dataset with different contrast phases predicting a partial nephrectomy all cortico-medullary phase dataset (0.67 ± 0.36 to 0.89 ± 0.10, 122/300 images flagged).</p><p><strong>Discussion: </strong>Comparing interfold sub-model disagreement against human interobserver values is an effective and efficient way to assess automated predictions when a reference standard is not available. This functionality provides a necessary safeguard to patient care important to safely implement automated medical image segmentation workflows.</p>\",\"PeriodicalId\":73101,\"journal\":{\"name\":\"Frontiers in radiology\",\"volume\":\"3 \",\"pages\":\"1223294\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-09-15\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC10540615/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Frontiers in radiology\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.3389/fradi.2023.1223294\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2023/1/1 0:00:00\",\"PubModel\":\"eCollection\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Frontiers in radiology","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3389/fradi.2023.1223294","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2023/1/1 0:00:00","PubModel":"eCollection","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

摘要

简介:为了将机器学习工作流程安全地实施到临床实践中,以及在模型训练过程中识别困难案例,迫切需要自动标记表现不佳的预测的方法。方法:使用折叠之间的骰子分数来量化五重交叉验证子模型之间的差异,并将其总结为模型置信度的替代品。将总结的折叠间骰子与由人类观察者间值通知的阈值进行比较,以确定是否应手动审查最终的集成模型性能。结果:该方法在所有任务中都有效地标记了较差的分割图像,而无需参考标准。使用中位数Interfold Dice进行比较,发现在排除标记图像后,域内CT(0.85±0.20至0.91±0.08,标记8/50图像)和MR(0.76±0.27至0.85±0.09,标记8/5图像)的骰子得分显著提高。最令人印象深刻的是,在模拟的分布外任务中,骰子得分有了显著的提高,在该任务中,模型在具有不同对比度阶段的根治性肾切除术数据集上进行训练,预测部分肾切除术全皮质-髓质阶段数据集(标记0.67±0.36至0.89±0.10122/300个图像)当没有参考标准时,评估自动预测的有效和高效的方法。该功能为患者护理提供了必要的保障,这对安全实施自动化医疗图像分割工作流程非常重要。
本文章由计算机程序翻译,如有差异,请以英文原文为准。

AI in the Loop: functionalizing fold performance disagreement to monitor automated medical image segmentation workflows.

AI in the Loop: functionalizing fold performance disagreement to monitor automated medical image segmentation workflows.

AI in the Loop: functionalizing fold performance disagreement to monitor automated medical image segmentation workflows.

AI in the Loop: functionalizing fold performance disagreement to monitor automated medical image segmentation workflows.

Introduction: Methods that automatically flag poor performing predictions are drastically needed to safely implement machine learning workflows into clinical practice as well as to identify difficult cases during model training.

Methods: Disagreement between the fivefold cross-validation sub-models was quantified using dice scores between folds and summarized as a surrogate for model confidence. The summarized Interfold Dices were compared with thresholds informed by human interobserver values to determine whether final ensemble model performance should be manually reviewed.

Results: The method on all tasks efficiently flagged poor segmented images without consulting a reference standard. Using the median Interfold Dice for comparison, substantial dice score improvements after excluding flagged images was noted for the in-domain CT (0.85 ± 0.20 to 0.91 ± 0.08, 8/50 images flagged) and MR (0.76 ± 0.27 to 0.85 ± 0.09, 8/50 images flagged). Most impressively, there were dramatic dice score improvements in the simulated out-of-distribution task where the model was trained on a radical nephrectomy dataset with different contrast phases predicting a partial nephrectomy all cortico-medullary phase dataset (0.67 ± 0.36 to 0.89 ± 0.10, 122/300 images flagged).

Discussion: Comparing interfold sub-model disagreement against human interobserver values is an effective and efficient way to assess automated predictions when a reference standard is not available. This functionality provides a necessary safeguard to patient care important to safely implement automated medical image segmentation workflows.

求助全文
通过发布文献求助,成功后即可免费获取论文全文。 去求助
来源期刊
CiteScore
1.20
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信