Tools and techniques for computational reproducibility | Publons

flavoursofopenscience's bookmarks 2018-09-08

Summary:

Abstract: When reporting research findings, scientists document the steps they followed so that others can verify and build upon the research. When those steps have been described in sufficient detail that others can retrace the steps and obtain similar results, the research is said to be reproducible. Computers play a vital role in many research disciplines and present both opportunities and challenges for reproducibility. Computers can be programmed to execute analysis tasks, and those programs can be repeated and shared with others. The deterministic nature of most computer programs means that the same analysis tasks, applied to the same data, will often produce the same outputs. However, in practice, computational findings often cannot be reproduced because of complexities in how software is packaged, installed, and executed—and because of limitations associated with how scientists document analysis steps. Many tools and techniques are available to help overcome these challenges; here we describe seven such strategies. With a broad scientific audience in mind, we describe the strengths and limitations of each approach, as well as the circumstances under which each might be applied. No single strategy is sufficient for every scenario; thus we emphasize that it is often useful to combine approaches.


Preprint & peer-reviews, published article at: https://dx.doi.org/10.1186/s13742-016-0135-4

Link:

https://publons.com/publon/781618/

Updated:

09/08/2018, 06:00

From feeds:

Open Access Tracking Project (OATP) » flavoursofopenscience's bookmarks

Tags:

oa.publons oa.peer_review oa.floss oa.reproducibility oa.analysis oa.outputs oa.tools

Authors:

Stephen R Piccolo, Michael B Frampton,

Copyright info:

CC BY 4.0

Date tagged:

09/08/2018, 10:00

Date published:

07/11/2016, 06:00