6 Evaluation of the repository
The code and related research artefacts in the original code repositories were evaluated against:
- The criteria of badges related to reproducibility from various organisations and journals.
- Recommendations from the STARS framework for the sharing of code and associated materials from discrete-event simulation models (Monks, Harper, and Mustafee (2024)).
Between each journal badge, there was often alot of overlap in criteria. Hence, a list of unique criteria was produced. The repositories are evaluated against this criteria, and then depending on which criteria they met, against the badges themselves.
Caveat: Please note that these criteria are based on available information about each badge online, and that we have likely differences in our procedure (e.g. allowed troubleshooting for execution and reproduction, not under tight time pressure to complete). Moreover, we focus only on reproduction of the discrete-event simulation, and not on other aspects of the article. We cannot guarantee that the badges below would have been awarded in practice by these journals.
6.1 Summary
Unique badge criteria:
Badges:
Essential components of STARS framework:
Optional components of STARS framework:
6.2 Journal badges
Key:
- S: Shoaib and Ramamohan (2021) - link to evaluation
- H: Huang et al. (2019) - link to evaluation
- L: Lim et al. (2020) - link to evaluation
- K: Kim et al. (2021) - link to evaluation
- A: Anagnostou et al. (2022) - link to evaluation
- J: Johnson et al. (2021) - link to evaluation
In this section and below, the criteria for each study are marked as either being fully met (✅), partially met (🟡), not met (❌) or not applicable (N/A).
Unique criteria:
Item | S | H | L | K | A | J | |
---|---|---|---|---|---|---|---|
Criteria related to how artefacts are shared | |||||||
Stored in a permanent archive that is publicly and openly accessible | ❌ | ❌ | ❌ | ❌ | ✅ | ||
Has a persistent identifier | ❌ | ❌ | ❌ | ❌ | ✅ | ||
Includes an open license | ❌ | ✅ | ❌ | ✅ | ✅ | ||
Criteria related to what artefacts are shared | |||||||
Artefacts are relevant to and contribute to the article’s results | ✅ | ✅ | ✅ | ✅ | ✅ | ||
Complete set of materials shared (as would be needed to fully reproduce article) | ❌ | ❌ | ❌ | ❌ | ❌ | ||
Criteria related to the structure and documentation of the artefacts | |||||||
Artefacts are well structured/organised (e.g. to the extent that reuse and repurposing is facilitated, adhering to norms and standards of research community) | ❌ | ❌ | ❌ | ✅ | ✅ | ||
Artefacts are sufficiently documented (i.e. to understand how it works, to enable it to be run, including package versions) | ❌ | ❌ | ❌ | ❌ | ✅ | ||
Artefacts are carefully documented (more than sufficient - i.e. to the extent that reuse and repurposing is facilitated - e.g. changing parameters, reusing for own purpose) | ❌ | ❌ | ❌ | ❌ | ✅ | ||
Artefacts are clearly documented and accompanied by a README file with step-by-step instructions on how to reproduce results in the manuscript | ❌ | ❌ | ❌ | ❌ | ✅ | ||
Criteria related to running and reproducing results | |||||||
Scripts can be successfully executed | ✅ | ✅ | ✅ | ✅ | ✅ | ||
Independent party regenerated results using the authors research artefacts | ✅ | ❌ | ✅ | ✅ | ✅ | ||
Reproduced within approximately one hour (excluding compute time) | ❌ | ❌ | ❌ | ❌ | ❌ | ❌ |
Badges:
The badges are grouped into three categories:
- “Open objects” badges: These badges relate to research artefacts being made openly available.
- “Object review” badges: These badges relate to the research artefacts being reviewed against criteria of the badge issuer.
- “Reproduced” badges: These badges relate to an independent party regenerating the reuslts of the article using the author objects.
Item | S | H | L | K | A | J | |
---|---|---|---|---|---|---|---|
“Open objects” badges | |||||||
NISO “Open Research Objects (ORO)” • Stored in a permanent archive that is publicly and openly accessible • Has a persistent identifier • Includes an open license |
❌ | ❌ | ❌ | ❌ | ✅ | ||
NISO “Open Research Objects - All (ORO-A)” • Stored in a permanent archive that is publicly and openly accessible • Has a persistent identifier • Includes an open license • Complete set of materials shared (as would be needed to fully reproduce article) |
❌ | ❌ | ❌ | ❌ | ❌ | ||
ACM “Artifacts Available” • Stored in a permanent archive that is publicly and openly accessible • Has a persistent identifier |
❌ | ❌ | ❌ | ❌ | ✅ | ||
COS “Open Code” • Stored in a permanent archive that is publicly and openly accessible • Has a persistent identifier • Includes an open license • Complete set of materials shared (as would be needed to fully reproduce article) • Artefacts are sufficiently documented (i.e. to understand how it works, to enable it to be run, including package versions) |
❌ | ❌ | ❌ | ❌ | ❌ | ||
IEEE “Code Available” • Complete set of materials shared (as would be needed to fully reproduce article) |
❌ | ❌ | ❌ | ❌ | ❌ | ||
“Object review” badges | |||||||
ACM “Artifacts Evaluated - Functional” • Artefacts are sufficiently documented (i.e. to understand how it works, to enable it to be run, including package versions) • Artefacts are relevant to and contribute to the article’s results • Complete set of materials shared (as would be needed to fully reproduce article) • Scripts can be successfully executed |
❌ | ❌ | ❌ | ❌ | ❌ | ||
ACM “Artifacts Evaluated - Reusable” • Artefacts are sufficiently documented (i.e. to understand how it works, to enable it to be run, including package versions) • Artefacts are carefully documented (more than sufficient - i.e. to the extent that reuse and repurposing is facilitated - e.g. changing parameters, reusing for own purpose) • Artefacts are relevant to and contribute to the article’s results • Complete set of materials shared (as would be needed to fully reproduce article) • Scripts can be successfully executed • Artefacts are well structured/organised (e.g. to the extent that reuse and repurposing is facilitated, adhering to norms and standards of research community) |
❌ | ❌ | ❌ | ❌ | ❌ | ||
IEEE “Code Reviewed” • Complete set of materials shared (as would be needed to fully reproduce article) • Scripts can be successfully executed |
❌ | ❌ | ❌ | ❌ | ❌ | ||
“Reproduced” badges | |||||||
NISO “Results Reproduced (ROR-R)” • Independent party regenerated results using the authors research artefacts |
✅ | ❌ | ✅ | ✅ | ✅ | ||
ACM “Results Reproduced” • Independent party regenerated results using the authors research artefacts |
✅ | ❌ | ✅ | ✅ | ✅ | ||
IEEE “Code Reproducible” • Independent party regenerated results using the authors research artefacts |
✅ | ❌ | ✅ | ✅ | ✅ | ||
Psychological Science “Computational Reproducibility” • Independent party regenerated results using the authors research artefacts • Reproduced within approximately one hour (excluding compute time) • Artefacts are well structured/organised (e.g. to the extent that reuse and repurposing is facilitated, adhering to norms and standards of research community) • Artefacts are clearly documented and accompanied by a README file with step-by-step instructions on how to reproduce results in the manuscript |
❌ | ❌ | ❌ | ❌ | ❌ |
6.3 STARS framework
Key:
- S: Shoaib and Ramamohan (2021) - link to evaluation
- H: Huang et al. (2019) - link to evaluation
- L: Lim et al. (2020) - link to evaluation
- K: Kim et al. (2021) - link to evaluation
- A: Anagnostou et al. (2022) - link to evaluation
- J: Johnson et al. (2021) - link to evaluation
Item | S | H | L | K | A | J | |
---|---|---|---|---|---|---|---|
Essential components | |||||||
Open license Free and open-source software (FOSS) license (e.g. MIT, GNU Public License (GPL)) |
❌ | ✅ | ❌ | ✅ | ✅ | ||
Dependency management Specify software libraries, version numbers and sources (e.g. dependency management tools like virtualenv, conda, poetry) |
❌ | ❌ | ❌ | 🟡 | ✅ | ||
FOSS model Coded in FOSS language (e.g. R, Julia, Python) |
✅ | ✅ | ✅ | ✅ | ✅ | ✅ | |
Minimum documentation Minimal instructions (e.g. in README) that overview (a) what model does, (b) how to install and run model to obtain results, and (c) how to vary parameters to run new experiments |
❌ | ❌ | ❌ | ✅ | ✅ | ||
ORCID ORCID for each study author |
❌ | ❌ | ❌ | ❌ | ❌ | ||
Citation information Instructions on how to cite the research artefact (e.g. CITATION.cff file) |
❌ | ❌ | ❌ | ❌ | ✅ | ||
Remote code repository Code available in a remote code repository (e.g. GitHub, GitLab, BitBucket) |
✅ | ✅ | ✅ | ✅ | ✅ | ||
Open science archive Code stored in an open science archive with FORCE11 compliant citation and guaranteed persistance of digital artefacts (e.g. Figshare, Zenodo, the Open Science Framework (OSF), and the Computational Modeling in the Social and Ecological Sciences Network (CoMSES Net)) |
❌ | ❌ | ❌ | ❌ | ✅ | ||
Optional components | |||||||
Enhanced documentation Open and high quality documentation on how the model is implemented and works (e.g. via notebooks and markdown files, brought together using software like Quarto and Jupyter Book). Suggested content includes: • Plain english summary of project and model • Clarifying license • Citation instructions • Contribution instructions • Model installation instructions • Structured code walk through of model • Documentation of modelling cycle using TRACE • Annotated simulation reporting guidelines • Clear description of model validation including its intended purpose |
❌ | ❌ | ❌ | ❌ | ❌ | ||
Documentation hosting Host documentation (e.g. with GitHub pages, GitLab pages, BitBucket Cloud, Quarto Pub) |
❌ | ❌ | ❌ | ❌ | ❌ | ||
Online coding environment Provide an online environment where users can run and change code (e.g. BinderHub, Google Colaboratory, Deepnote) |
❌ | ❌ | ❌ | ❌ | ❌ | ||
Model interface Provide web application interface to the model so it is accessible to less technical simulation users |
❌ | ✅ | ❌ | ❌ | ✅ | ||
Web app hosting Host web app online (e.g. Streamlit Community Cloud, ShinyApps hosting) |
❌ | ✅ | ❌ | ❌ | 🟡 |
6.4 Timings
6.5 Badge sources
National Information Standards Organisation (NISO) (NISO Reproducibility Badging and Definitions Working Group (2021))
- “Open Research Objects (ORO)”
- “Open Research Objects - All (ORO-A)”
- “Results Reproduced (ROR-R)”
Association for Computing Machinery (ACM) (Association for Computing Machinery (ACM) (2020))
- “Artifacts Available”
- “Artifacts Evaluated - Functional”
- “Artifacts Evaluated - Resuable”
- “Results Reproduced”
Center for Open Science (COS) (Blohowiak et al. (2023))
- “Open Code”
Institute of Electrical and Electronics Engineers (IEEE) (Institute of Electrical and Electronics Engineers (IEEE) (n.d.))
- “Code Available”
- “Code Reviewed”
- “Code Reproducible”
Psychological Science (Hardwicke and Vazire (2023) and Association for Psychological Science (APS) (2023))
- “Computational Reproducibility”