Comparing perturbation models for evaluating stability of neuroimaging pipelines.

Gregory Kiar, Pablo de Oliveira Castro, Pierre Rioux, Eric Petit, Shawn T Brown, Alan C Evans, Tristan Glatard
{"title":"Comparing perturbation models for evaluating stability of neuroimaging pipelines.","authors":"Gregory Kiar,&nbsp;Pablo de Oliveira Castro,&nbsp;Pierre Rioux,&nbsp;Eric Petit,&nbsp;Shawn T Brown,&nbsp;Alan C Evans,&nbsp;Tristan Glatard","doi":"10.1177/1094342020926237","DOIUrl":null,"url":null,"abstract":"<p><p>With an increase in awareness regarding a troubling lack of reproducibility in analytical software tools, the degree of validity in scientific derivatives and their downstream results has become unclear. The nature of reproducibility issues may vary across domains, tools, data sets, and computational infrastructures, but numerical instabilities are thought to be a core contributor. In neuroimaging, unexpected deviations have been observed when varying operating systems, software implementations, or adding negligible quantities of noise. In the field of numerical analysis, these issues have recently been explored through Monte Carlo Arithmetic, a method involving the instrumentation of floating-point operations with probabilistic noise injections at a target precision. Exploring multiple simulations in this context allows the characterization of the result space for a given tool or operation. In this article, we compare various perturbation models to introduce instabilities within a typical neuroimaging pipeline, including (i) targeted noise, (ii) Monte Carlo Arithmetic, and (iii) operating system variation, to identify the significance and quality of their impact on the resulting derivatives. We demonstrate that even low-order models in neuroimaging such as the structural connectome estimation pipeline evaluated here are sensitive to numerical instabilities, suggesting that stability is a relevant axis upon which tools are compared, alongside more traditional criteria such as biological feasibility, computational efficiency, or, when possible, accuracy. Heterogeneity was observed across participants which clearly illustrates a strong interaction between the tool and data set being processed, requiring that the stability of a given tool be evaluated with respect to a given cohort. We identify use cases for each perturbation method tested, including quality assurance, pipeline error detection, and local sensitivity analysis, and make recommendations for the evaluation of stability in a practical and analytically focused setting. Identifying how these relationships and recommendations scale to higher order computational tools, distinct data sets, and their implication on biological feasibility remain exciting avenues for future work.</p>","PeriodicalId":506320,"journal":{"name":"The International Journal of High Performance Computing Applications","volume":"34 5","pages":"491-501"},"PeriodicalIF":0.0000,"publicationDate":"2020-09-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://sci-hub-pdf.com/10.1177/1094342020926237","citationCount":"11","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"The International Journal of High Performance Computing Applications","FirstCategoryId":"94","ListUrlMain":"https://doi.org/10.1177/1094342020926237","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2020/5/21 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 11

Abstract

With an increase in awareness regarding a troubling lack of reproducibility in analytical software tools, the degree of validity in scientific derivatives and their downstream results has become unclear. The nature of reproducibility issues may vary across domains, tools, data sets, and computational infrastructures, but numerical instabilities are thought to be a core contributor. In neuroimaging, unexpected deviations have been observed when varying operating systems, software implementations, or adding negligible quantities of noise. In the field of numerical analysis, these issues have recently been explored through Monte Carlo Arithmetic, a method involving the instrumentation of floating-point operations with probabilistic noise injections at a target precision. Exploring multiple simulations in this context allows the characterization of the result space for a given tool or operation. In this article, we compare various perturbation models to introduce instabilities within a typical neuroimaging pipeline, including (i) targeted noise, (ii) Monte Carlo Arithmetic, and (iii) operating system variation, to identify the significance and quality of their impact on the resulting derivatives. We demonstrate that even low-order models in neuroimaging such as the structural connectome estimation pipeline evaluated here are sensitive to numerical instabilities, suggesting that stability is a relevant axis upon which tools are compared, alongside more traditional criteria such as biological feasibility, computational efficiency, or, when possible, accuracy. Heterogeneity was observed across participants which clearly illustrates a strong interaction between the tool and data set being processed, requiring that the stability of a given tool be evaluated with respect to a given cohort. We identify use cases for each perturbation method tested, including quality assurance, pipeline error detection, and local sensitivity analysis, and make recommendations for the evaluation of stability in a practical and analytically focused setting. Identifying how these relationships and recommendations scale to higher order computational tools, distinct data sets, and their implication on biological feasibility remain exciting avenues for future work.

Abstract Image

Abstract Image

Abstract Image

比较摄动模型评价神经成像管道的稳定性。
随着对分析软件工具中令人不安的再现性缺乏的认识的增加,科学衍生物及其下游结果的有效性程度已变得不清楚。可再现性问题的性质可能因领域、工具、数据集和计算基础设施而异,但数值不稳定性被认为是一个核心因素。在神经成像中,当不同的操作系统、软件实现或添加可忽略不计的噪声时,会观察到意想不到的偏差。在数值分析领域,这些问题最近通过蒙特卡罗算法进行了探索,蒙特卡罗算法是一种涉及在目标精度上注入概率噪声的浮点运算仪器的方法。在这种情况下探索多个模拟允许对给定工具或操作的结果空间进行表征。在本文中,我们比较了各种扰动模型,以引入典型神经成像管道中的不稳定性,包括(i)目标噪声,(ii)蒙特卡罗算法和(iii)操作系统变化,以确定它们对所得导数影响的重要性和质量。我们证明,即使是神经成像中的低阶模型,如本文评估的结构连接体估计管道,对数值不稳定性也很敏感,这表明稳定性是一个相关的轴,在这个轴上比较工具,以及更传统的标准,如生物可行性、计算效率,或者在可能的情况下,准确性。在参与者中观察到异质性,这清楚地说明了工具和正在处理的数据集之间的强交互作用,要求对给定的队列评估给定工具的稳定性。我们为测试的每种扰动方法确定用例,包括质量保证、管道错误检测和局部敏感性分析,并在实际和分析集中的环境中对稳定性评估提出建议。确定这些关系和建议如何扩展到更高阶的计算工具、不同的数据集,以及它们对生物学可行性的影响,仍然是未来工作的令人兴奋的途径。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信