Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems最新文献

筛选
英文 中文
Improving Reproducibility of Distributed Computational Experiments 提高分布式计算实验的再现性
Quan Pham, T. Malik, Dai Hai Ton That, A. Youngdahl
{"title":"Improving Reproducibility of Distributed Computational Experiments","authors":"Quan Pham, T. Malik, Dai Hai Ton That, A. Youngdahl","doi":"10.1145/3214239.3214241","DOIUrl":"https://doi.org/10.1145/3214239.3214241","url":null,"abstract":"Conference and journal publications increasingly require experiments associated with a submitted article to be repeatable. Authors comply to this requirement by sharing all associated digital artifacts, i.e., code, data, and environment configuration scripts. To ease aggregation of the digital artifacts, several tools have recently emerged that automate the aggregation of digital artifacts by auditing an experiment execution and building a portable container of code, data, and environment. However, current tools only package non-distributed computational experiments. Distributed computational experiments must either be packaged manually or supplemented with sufficient documentation. In this paper, we outline the reproducibility requirements of distributed experiments using a distributed computational science experiment involving use of message-passing interface (MPI), and propose a general method for auditing and repeating distributed experiments. Using Sciunit we show how this method can be implemented. We validate our method with initial experiments showing application re-execution runtime can be improved by 63% with a trade-off of longer run-time on initial audit execution.","PeriodicalId":422030,"journal":{"name":"Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems","volume":"92 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132566099","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems 第一届计算机系统实用可重复评估国际研讨会论文集
{"title":"Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems","authors":"","doi":"10.1145/3214239","DOIUrl":"https://doi.org/10.1145/3214239","url":null,"abstract":"","PeriodicalId":422030,"journal":{"name":"Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems","volume":"2 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134488567","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Popper Pitfalls: Experiences Following a Reproducibility Convention 波普尔陷阱:遵循可重复性惯例的经验
Michael Sevilla, C. Maltzahn
{"title":"Popper Pitfalls: Experiences Following a Reproducibility Convention","authors":"Michael Sevilla, C. Maltzahn","doi":"10.1145/3214239.3214243","DOIUrl":"https://doi.org/10.1145/3214239.3214243","url":null,"abstract":"We describe the four publications we have tried to make reproducible and discuss how each paper has changed our workflows, practices, and collaboration policies. The fundamental insight is that paper artifacts must be made reproducible from the start of the project; artifacts are too difficult to make reproducible when the papers are (1) already published and (2) authored by researchers that are not thinking about reproducibility. In this paper, we present the best practices adopted by our research laboratory, which was sculpted by the pitfalls we have identified for the Popper convention. We conclude with a \"call-to-arms\" for the community focused on enhancing reproducibility initiatives for academic conferences, industry environments, and national laboratories. We hope that our experiences will shape a best practices guide for future reproducible papers.","PeriodicalId":422030,"journal":{"name":"Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems","volume":"3 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128886466","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 3
Software Provenance: Track the Reality Not the Virtual Machine 软件来源:跟踪现实而不是虚拟机
D. Wilkinson, Luis Oliveira, D. Mossé, B. Childers
{"title":"Software Provenance: Track the Reality Not the Virtual Machine","authors":"D. Wilkinson, Luis Oliveira, D. Mossé, B. Childers","doi":"10.1145/3214239.3214244","DOIUrl":"https://doi.org/10.1145/3214239.3214244","url":null,"abstract":"The growing use of computers and massive storage by individuals is driving interest in digital preservation. The scientific method demands accountability through digital reproducibility, adding another strong motivation for preservation. However, data alone can become obsolete if the interactivity of software required to interpret the data is lost. Virtual machines (VMs) may preserve interactivity however do so at the cost of obscuring the nature of what lies within. Occam, instead, builds VMs on-the-fly while storing and distributing well-described software packages. Thus, the system can track the exact components inside VMs without storing the machines themselves, allowing software to be repeatably built and executed. For Occam to recreate VMs, it needs to know exactly what software was used within. Through this tracking, such software can even be modified and rebuilt. Occam keeps track of all such components in manifests, allowing anybody to know exactly what is in each VM, and the origins of each component.","PeriodicalId":422030,"journal":{"name":"Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124397367","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 5
Supporting Long-term Reproducible Software Execution 支持长期可复制的软件执行
Luis Oliveira, D. Wilkinson, D. Mossé, B. Childers
{"title":"Supporting Long-term Reproducible Software Execution","authors":"Luis Oliveira, D. Wilkinson, D. Mossé, B. Childers","doi":"10.1145/3214239.3214245","DOIUrl":"https://doi.org/10.1145/3214239.3214245","url":null,"abstract":"A recent widespread realization that software experiments are not as easily replicated as once believed brought software execution preservation to the science spotlight. As a result, scientists, institutions, and funding agencies have recently been pushing for the development of methodologies and tools that preserve software artifacts. Despite current efforts, long term reproducibility still eludes us. In this paper, we present the requirements for software execution preservation and discuss how to improve long-term reproducibility in science. In particular, we discuss the reasons why preserving binaries and pre-built execution environments is not enough and why preserving the ability to replicate results is not the same as preserving software for reproducible science. Finally, we show how these requirements are supported by Occam, an open curation framework that fully preserves software and its dependencies from source to execution, promoting transparency, longevity, and re-use. Specifically, Occam provides the ability to automatically deploy workflows in a fully-functional environment that is able to not only run them, but make them easily replicable.","PeriodicalId":422030,"journal":{"name":"Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems","volume":"2014 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121594954","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 11
Enabling the Verification of Computational Results: An Empirical Evaluation of Computational Reproducibility 使计算结果的验证:计算再现性的经验评价
V. Stodden, M. Krafczyk, A. Bhaskar
{"title":"Enabling the Verification of Computational Results: An Empirical Evaluation of Computational Reproducibility","authors":"V. Stodden, M. Krafczyk, A. Bhaskar","doi":"10.1145/3214239.3214242","DOIUrl":"https://doi.org/10.1145/3214239.3214242","url":null,"abstract":"The ability to independently regenerate published computational claims is widely recognized as a key component of scientific reproducibility. In this article we take a narrow interpretation of this goal, and attempt to regenerate published claims from author-supplied information, including data, code, inputs, and other provided specifications, on a different computational system than that used by the original authors. We are motivated by Claerbout and Donoho's exhortation of the importance of providing complete information for reproducibility of the published claim. We chose the Elsevier journal, the Journal of Computational Physics, which has stated author guidelines that encourage the availability of computational digital artifacts that support scholarly findings. In an IRB approved study at the University of Illinois at Urbana-Champaign (IRB #17329) we gathered artifacts from a sample of authors who published in this journal in 2016 and 2017. We then used the ICERM criteria generated at the 2012 ICERM workshop \"Reproducibility in Computational and Experimental Mathematics\" to evaluate the sufficiency of the information provided in the publications and the ease with which the digital artifacts afforded computational reproducibility. We find that, for the articles for which we obtained computational artifacts, we could not easily regenerate the findings for 67% of them, and we were unable to easily regenerate all the findings for any of the articles. We then evaluated the artifacts we did obtain (55 of 306 articles) and find that the main barriers to computational reproducibility are inadequate documentation of code, data, and workflow information (70.9%), missing code function and setting information, and missing licensing information (75%). We recommend improvements based on these findings, including the deposit of supporting digital artifacts for reproducibility as a condition of publication, and verification of computational findings via re-execution of the code when possible.","PeriodicalId":422030,"journal":{"name":"Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123427160","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 20
MonEx: An Integrated Experiment Monitoring Framework Standing on Off-The-Shelf Components MonEx:基于现成组件的集成实验监控框架
Abdulqawi Saif, Alexandre Merlin, L. Nussbaum, Yeqiong Song
{"title":"MonEx: An Integrated Experiment Monitoring Framework Standing on Off-The-Shelf Components","authors":"Abdulqawi Saif, Alexandre Merlin, L. Nussbaum, Yeqiong Song","doi":"10.1145/3214239.3214240","DOIUrl":"https://doi.org/10.1145/3214239.3214240","url":null,"abstract":"Most computer experiments include a phase where metrics are gathered from and about various kinds of resources. This phase is often done via manual, non-reproducible and error-prone steps. Infrastructure monitoring tools facilitate collecting experiments' data to some extent. However, there is no conventional way for doing so, and there is still much work to be done (e.g. capturing user experiments) to leverage the monitoring activity for monitoring experiments. To overcome those challenges, we define the requirements of experiments monitoring, clarifying Experiment Monitoring Frameworks' scope and mainly focusing on reusability of experiments' data, and portability of experiments' metrics. We then propose MonEx EMF that satisfies those requirements. MonEx is built on top of infrastructure monitoring solutions and supports various monitoring approaches. It fully integrates into the experiment workflow by encompassing all steps from data acquisition to producing publishable figures. Hence, MonEx represents a first step towards unifying methods of collecting experiments' data.","PeriodicalId":422030,"journal":{"name":"Proceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems","volume":"27 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2018-06-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121592342","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信