Visible to the public Improving Reproducibility of Distributed Computational Experiments

TitleImproving Reproducibility of Distributed Computational Experiments
Publication TypeConference Paper
Year of Publication2018
AuthorsPham, Quan, Malik, Tanu, That, Dai Hai Ton, Youngdahl, Andrew
Conference NameProceedings of the First International Workshop on Practical Reproducible Evaluation of Computer Systems
PublisherACM
Conference LocationNew York, NY, USA
ISBN Number978-1-4503-5861-3
Keywordscomposability, Human Behavior, Metrics, Network provenance, Provenance, pubcrawl, record and replay, reproducibility of distributed objects, Resiliency, sciunit
AbstractConference and journal publications increasingly require experiments associated with a submitted article to be repeatable. Authors comply to this requirement by sharing all associated digital artifacts, i.e., code, data, and environment configuration scripts. To ease aggregation of the digital artifacts, several tools have recently emerged that automate the aggregation of digital artifacts by auditing an experiment execution and building a portable container of code, data, and environment. However, current tools only package non-distributed computational experiments. Distributed computational experiments must either be packaged manually or supplemented with sufficient documentation. In this paper, we outline the reproducibility requirements of distributed experiments using a distributed computational science experiment involving use of message-passing interface (MPI), and propose a general method for auditing and repeating distributed experiments. Using Sciunit we show how this method can be implemented. We validate our method with initial experiments showing application re-execution runtime can be improved by 63% with a trade-off of longer run-time on initial audit execution.
URLhttp://doi.acm.org/10.1145/3214239.3214241
DOI10.1145/3214239.3214241
Citation Keypham_improving_2018