Proceedings of the 2023 ACM Conference on Reproducibility and Replicability最新文献

筛选
英文 中文
We Need More Reproducibility Content Across the Computer Science Curriculum 我们需要在整个计算机科学课程中增加可重复性内容
Proceedings of the 2023 ACM Conference on Reproducibility and Replicability Pub Date : 2023-06-27 DOI: 10.1145/3589806.3600033
Fraida Fund
{"title":"We Need More Reproducibility Content Across the Computer Science Curriculum","authors":"Fraida Fund","doi":"10.1145/3589806.3600033","DOIUrl":"https://doi.org/10.1145/3589806.3600033","url":null,"abstract":"With increasing recognition of the importance of reproducibility in computer science research, a wide range of efforts to promote reproducible research have been implemented across various sub-disciplines of computer science. These include artifact review and badging processes, and dedicated reproducibility tracks at conferences. However, these initiatives primarily engage active researchers and students already involved in research in their respective areas. In this paper, we present an argument for expanding the scope of these efforts to include a much larger audience, by introducing more reproducibility content into computer science courses. We describe various ways to integrate reproducibility content into the curriculum, drawing on our own experiences, as well as published experience reports from several sub-disciplines of computer science and computational science.","PeriodicalId":393751,"journal":{"name":"Proceedings of the 2023 ACM Conference on Reproducibility and Replicability","volume":"25 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129606486","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
On Reporting Robust and Trustworthy Conclusions from Model Comparison Studies Involving Neural Networks and Randomness 从涉及神经网络和随机性的模型比较研究中报告稳健和可信的结论
Proceedings of the 2023 ACM Conference on Reproducibility and Replicability Pub Date : 2023-06-27 DOI: 10.1145/3589806.3600044
Odd Erik Gundersen, Saeid Shamsaliei, H. S. Kjærnli, H. Langseth
{"title":"On Reporting Robust and Trustworthy Conclusions from Model Comparison Studies Involving Neural Networks and Randomness","authors":"Odd Erik Gundersen, Saeid Shamsaliei, H. S. Kjærnli, H. Langseth","doi":"10.1145/3589806.3600044","DOIUrl":"https://doi.org/10.1145/3589806.3600044","url":null,"abstract":"The performance of neural networks differ when the only difference is the seed initializing the pseudo-random number generator that generates random numbers for their training. In this paper we are concerned with how random initialization affect the conclusions that we draw from experiments with neural networks. We run a high number of repeated experiments using state of the art models for time-series prediction and image classification to investigate this statistical phenomenon. Our investigations show that erroneous conclusions can easily be drawn from such experiments. Based on these observations we propose several measures that will improve the robustness and trustworthiness of conclusions inferred from model comparison studies with small absolute effect sizes.","PeriodicalId":393751,"journal":{"name":"Proceedings of the 2023 ACM Conference on Reproducibility and Replicability","volume":"14 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126482158","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Towards Reproducible Execution of Closed-Source Applications from Internet Archives 从互联网档案中实现闭源应用程序的可重复执行
Proceedings of the 2023 ACM Conference on Reproducibility and Replicability Pub Date : 2023-06-27 DOI: 10.1145/3589806.3600035
M. Satyanarayanan, J. Harkes, J. Blakley
{"title":"Towards Reproducible Execution of Closed-Source Applications from Internet Archives","authors":"M. Satyanarayanan, J. Harkes, J. Blakley","doi":"10.1145/3589806.3600035","DOIUrl":"https://doi.org/10.1145/3589806.3600035","url":null,"abstract":"Olive enables execution of closed-source applications decades after their creation. With appropriate authentication and authorization, anyone on the Internet can execute any archived application with no more effort than a mouse click. User experience is good, even for an interaction-intensive application. Olive uses virtual machine (VM) technology to encapsulate legacy software, including the operating system and all layers above it. If the legacy hardware is already obsolete at curation time, an emulator for it on more modern hardware can be included within the VM image. This paper is an experience report on the decade-long evolution of this concept.","PeriodicalId":393751,"journal":{"name":"Proceedings of the 2023 ACM Conference on Reproducibility and Replicability","volume":"17 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"115076966","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Automatic Reproduction of Workflows in the Snakemake Workflow Catalog and nf-core Registries 在snakemaker工作流目录和非核心注册表中自动复制工作流
Proceedings of the 2023 ACM Conference on Reproducibility and Replicability Pub Date : 2023-06-27 DOI: 10.1145/3589806.3600037
Samuel Grayson, D. Marinov, Daniel S. Katz, Reed Milewicz
{"title":"Automatic Reproduction of Workflows in the Snakemake Workflow Catalog and nf-core Registries","authors":"Samuel Grayson, D. Marinov, Daniel S. Katz, Reed Milewicz","doi":"10.1145/3589806.3600037","DOIUrl":"https://doi.org/10.1145/3589806.3600037","url":null,"abstract":"Workflows make it easier for scientists to assemble computational experiments consisting of many disparate components. However, those disparate components also increase the probability that the computational experiment fails to be reproducible. Even if software is reproducible today, it may become irreproducible tomorrow without the software itself changing at all, because of the constantly changing software environment in which the software is run. To alleviate irreproducibility, workflow engines integrate with container engines. Additionally, communities that sprung up around workflow engines started to host registries for workflows that follow standards. These standards reduce the effort needed to make workflows automatically reproducible. In this paper, we study automatic reproduction of workflows from two registries, focusing on non-crashing executions. The experimental data lets us analyze the upper bound to which workflow engines could achieve reproducibility. We identify lessons learned in achieving reproducibility in practice.","PeriodicalId":393751,"journal":{"name":"Proceedings of the 2023 ACM Conference on Reproducibility and Replicability","volume":"20 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"117140389","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
KheOps: Cost-effective Repeatability, Reproducibility, and Replicability of Edge-to-Cloud Experiments KheOps:具有成本效益的可重复性、再现性和边缘到云实验的可复制性
Proceedings of the 2023 ACM Conference on Reproducibility and Replicability Pub Date : 2023-06-27 DOI: 10.1145/3589806.3600032
Daniel Rosendo, K. Keahey, Alexandru Costan, Matthieu Simonin, P. Valduriez, Gabriel Antoniu
{"title":"KheOps: Cost-effective Repeatability, Reproducibility, and Replicability of Edge-to-Cloud Experiments","authors":"Daniel Rosendo, K. Keahey, Alexandru Costan, Matthieu Simonin, P. Valduriez, Gabriel Antoniu","doi":"10.1145/3589806.3600032","DOIUrl":"https://doi.org/10.1145/3589806.3600032","url":null,"abstract":"Distributed infrastructures for computation and analytics are now evolving towards an interconnected ecosystem allowing complex scientific workflows to be executed across hybrid systems spanning from IoT Edge devices to Clouds, and sometimes to supercomputers (the Computing Continuum). Understanding the performance trade-offs of large-scale workflows deployed on such complex Edge-to-Cloud Continuum is challenging. To achieve this, one needs to systematically perform experiments, to enable their reproducibility and allow other researchers to replicate the study and the obtained conclusions on different infrastructures. This breaks down to the tedious process of reconciling the numerous experimental requirements and constraints with low-level infrastructure design choices. To address the limitations of the main state-of-the-art approaches for distributed, collaborative experimentation, such as Google Colab, Kaggle, and Code Ocean, we propose KheOps, a collaborative environment specifically designed to enable cost-effective reproducibility and replicability of Edge-to-Cloud experiments. KheOps is composed of three core elements: (1) an experiment repository; (2) a notebook environment; and (3) a multi-platform experiment methodology. We illustrate KheOps with a real-life Edge-to-Cloud application. The evaluations explore the point of view of the authors of an experiment described in an article (who aim to make their experiments reproducible) and the perspective of their readers (who aim to replicate the experiment). The results show how KheOps helps authors to systematically perform repeatable and reproducible experiments on the Grid5000 + FIT IoT LAB testbeds. Furthermore, KheOps helps readers to cost-effectively replicate authors experiments in different infrastructures such as Chameleon Cloud + CHI@Edge testbeds, and obtain the same conclusions with high accuracies (> 88% for all performance metrics).","PeriodicalId":393751,"journal":{"name":"Proceedings of the 2023 ACM Conference on Reproducibility and Replicability","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122507641","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Towards Evidence-Based Software Quality Practices for Reproducibility: Preliminary Results and Research Directions 面向可重复性的循证软件质量实践:初步结果和研究方向
Proceedings of the 2023 ACM Conference on Reproducibility and Replicability Pub Date : 2023-06-27 DOI: 10.1145/3589806.3600040
Reed Milewicz, Miranda R. Mundt
{"title":"Towards Evidence-Based Software Quality Practices for Reproducibility: Preliminary Results and Research Directions","authors":"Reed Milewicz, Miranda R. Mundt","doi":"10.1145/3589806.3600040","DOIUrl":"https://doi.org/10.1145/3589806.3600040","url":null,"abstract":"In the computational science and engineering (CSE) community, there is a prevailing belief that adopting better software development practices and investing in software quality will directly lead to more robust, reproducible software. There is, however, relatively little evidence to indicate what specific aspects of software quality influence reproducibility or which practices lead to more reproducible software. To better inform this discussion, we present preliminary findings from an ongoing study of how software quality practices among CSE teams affect the reproducibility of the software they create.","PeriodicalId":393751,"journal":{"name":"Proceedings of the 2023 ACM Conference on Reproducibility and Replicability","volume":"30 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114621333","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
GTP Benchmarks for Gradual Typing Performance 渐进式输入性能的GTP基准测试
Proceedings of the 2023 ACM Conference on Reproducibility and Replicability Pub Date : 2023-06-27 DOI: 10.1145/3589806.3600034
B. Greenman
{"title":"GTP Benchmarks for Gradual Typing Performance","authors":"B. Greenman","doi":"10.1145/3589806.3600034","DOIUrl":"https://doi.org/10.1145/3589806.3600034","url":null,"abstract":"Reproducible, rigorous experiments are key to effective computing research because they provide grounding and a way to measure progress. Gradual typing is an emerging area that desperately needs such grounding. A gradual language lets programmers add types to part of a codebase while leaving the rest untyped. The critical research question is how to balance the guarantees that types provide against the run-time cost of enforcing them. Either weaker guarantees or better implementation methods could lead to answers, but without benchmarks for reproducibility there is no sound way to evaluate competing designs. The GTP Benchmark Suite is a rigorous testbed for gradual typing that supports reproducible experiments. Starting from a core suite of 21 programs drawn from a variety of applications, it enables the systematic exploration of over 40K gradually-typed program configurations via software for managing experiments and for analyzing results. Language designers have used the benchmarks to evaluate implementation strategies in at least seven major efforts since 2014. Furthermore, the benchmarks have proven useful for broader topics in gradual typing.","PeriodicalId":393751,"journal":{"name":"Proceedings of the 2023 ACM Conference on Reproducibility and Replicability","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123148657","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
A Siren Song of Open Source Reproducibility, Examples from Machine Learning 开源再现性的警笛之歌,来自机器学习的例子
Proceedings of the 2023 ACM Conference on Reproducibility and Replicability Pub Date : 2023-06-27 DOI: 10.1145/3589806.3600042
Edward Raff, Andrew L. Farris
{"title":"A Siren Song of Open Source Reproducibility, Examples from Machine Learning","authors":"Edward Raff, Andrew L. Farris","doi":"10.1145/3589806.3600042","DOIUrl":"https://doi.org/10.1145/3589806.3600042","url":null,"abstract":"As reproducibility becomes a greater concern, conferences have largely converged to a strategy of asking reviewers to indicate whether code was attached to a submission. This represents a broader pattern of implementing actions based on presumed ideals, without studying whether those actions will produce positive results. We argue that focusing on code as a means of reproduction is misguided if we want to improve the state of reproducible and replicable research. In this study, we find this focus on code may be harmful — we should not force code to be submitted. Furthermore, there is a lack of evidence that conferences take effective actions to encourage and reward reproducibility. We argue that venues must take more action to advance reproducible machine learning research today.","PeriodicalId":393751,"journal":{"name":"Proceedings of the 2023 ACM Conference on Reproducibility and Replicability","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"125294517","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Integrated Reproducibility with Self-describing Machine Learning Models 集成再现性与自描述机器学习模型
Proceedings of the 2023 ACM Conference on Reproducibility and Replicability Pub Date : 2023-06-27 DOI: 10.1145/3589806.3600039
J. Wonsil, J. Sullivan, Margo Seltzer, A. Pocock
{"title":"Integrated Reproducibility with Self-describing Machine Learning Models","authors":"J. Wonsil, J. Sullivan, Margo Seltzer, A. Pocock","doi":"10.1145/3589806.3600039","DOIUrl":"https://doi.org/10.1145/3589806.3600039","url":null,"abstract":"Researchers and data scientists frequently want to collaborate on machine learning models. However, in the presence of sharing and simultaneous experimentation, it is challenging both to determine if two models were trained identically and to reproduce precisely someone else’s training process. We demonstrate how provenance collection that is tightly integrated into a machine learning library facilitates reproducibility. We present MERIT, a reproducibility system that leverages a robust configuration system and extensive provenance collection to exactly reproduce models, given only a model object. We integrate MERIT with Tribuo, an open-source Java-based machine learning library. Key features of this integrated reproducibility framework include controlling for sources of non-determinism in a multi-threaded environment and exposing the training differences between two models in a human-readable form. Our system allows simple reproduction of deployed Tribuo models without any additional information, ensuring data science research is reproducible. Our framework is open-source and available under an Apache 2.0 license.","PeriodicalId":393751,"journal":{"name":"Proceedings of the 2023 ACM Conference on Reproducibility and Replicability","volume":"149 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127262325","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Fingerprinting and Building Large Reproducible Datasets 指纹识别和构建大型可重复数据集
Proceedings of the 2023 ACM Conference on Reproducibility and Replicability Pub Date : 2023-06-20 DOI: 10.1145/3589806.3600043
Romain Lefeuvre, Jessie Galasso, B. Combemale, H. Sahraoui, Stefano Zacchiroli
{"title":"Fingerprinting and Building Large Reproducible Datasets","authors":"Romain Lefeuvre, Jessie Galasso, B. Combemale, H. Sahraoui, Stefano Zacchiroli","doi":"10.1145/3589806.3600043","DOIUrl":"https://doi.org/10.1145/3589806.3600043","url":null,"abstract":"Obtaining a relevant dataset is central to conducting empirical studies in software engineering. However, in the context of mining software repositories, the lack of appropriate tooling for large scale mining tasks hinders the creation of new datasets. Moreover, limitations related to data sources that change over time (e.g., code bases) and the lack of documentation of extraction processes make it difficult to reproduce datasets over time. This threatens the quality and reproducibility of empirical studies. In this paper, we propose a tool-supported approach facilitating the creation of large tailored datasets while ensuring their reproducibility. We leveraged all the sources feeding the Software Heritage append-only archive which are accessible through a unified programming interface to outline a reproducible and generic extraction process. We propose a way to define a unique fingerprint to characterize a dataset which, when provided to the extraction process, ensures that the same dataset will be extracted. We demonstrate the feasibility of our approach by implementing a prototype. We show how it can help reduce the limitations researchers face when creating or reproducing datasets.","PeriodicalId":393751,"journal":{"name":"Proceedings of the 2023 ACM Conference on Reproducibility and Replicability","volume":"242 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2023-06-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"131118406","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信