Ongoing technological developments have made it easier than ever before for scientists to share their data, materials, and analysis code. Sharing data and analysis code makes it easier for other researchers to re-use or check published research. These benefits will only emerge if researchers can reproduce the analysis reported in published articles, and if data is annotated well enough so that it is clear what all variables mean. Because most researchers have not been trained in computational reproducibility, it is important to evaluate current practices to identify practices that can be improved. We examined data and code sharing, as well as computational reproducibility of the main results, without contacting the original authors, for Registered Reports published in the psychological literature between 2014 and 2018. Of the 62 articles that met our inclusion criteria, data was available for 40 articles, and analysis scripts for 37 articles. For the 35 articles that shared both data and code and performed analyses in SPSS, R, Python, MATLAB, or JASP, we could run the scripts for 31 articles, and reproduce the main results for 20 articles. Although the articles that shared both data and code (35 out of 62, or 56%) and articles that could be computationally reproduced (20 out of 35, or 57%) was relatively high compared to other studies, there is clear room for improvement. We provide practical recommendations based on our observations, and link to examples of good research practices in the papers we reproduced.
An ecosystem of free open source tools for improving the rigor and reproducibility of research is thriving. Information professionals at research institutions must stay informed about what tools are available and how they compare. Ideally, information professionals can also onboard researchers to kickstart adoption of these tools. However, developing quality curriculum to train researchers on new tools requires expertise in the tool itself, which leaves many researchers without training on tools that may benefit their research.
This course will train participants to run hands-on, quality modules designed to onboard researchers to four free open source tools. Participants will experience each module, practice the exercises, and explore the training material needed to run the module themselves. An instructor guide that includes the module outline, objectives, description, frequently asked questions, pre- and post-participant surveys, target audience, and instructions for running a successful module is provided for each tool taught.
This course will train participants to run modules on unique aspects of four free open source tools for researchers:
Binder: Share your computational environment, code, and research notebooks.
Renku: Document and share your analysis pipelines.
Open Science Framework: Create a centralized, structured workspace for your research materials.
KnitR: Knit your R code with your analysis narrative in one executable research notebook and capture your dependencies.
Many participants already run short-duration training events at their institutions. This course is ideal for those participants who wish to improve the quality and variety of the training they already offer to researchers. Participants who do not currently run short-duration training events at their institutions will benefit from the course by learning an accessible and efficient way of getting started with these four modules.