Specification of Assessment Code Requirements - Task 002
Outcomes
- We have defined recommended checklists for submission which can be found in Appendix B of the performance assessment workbook.
- This recommended checklist then defines the Microsoft Form which is used as the interface for submitting codes to SHAREing.
- Collectively the checklist in the guidebook, the benchmarking guidelines and Microsoft Form consitute the submission interface and can be here or under the ‘Performance Assessment’ tab.
Fit to programme
This task has been identified by the working groups as part of the agenda behind WP 1.1.
The task number is 002.
Description
SHAREing wants to enable RTP groups to build up a performance assessment service: Groups can submit their codes and get a high-level overview back, from which we can identify next development steps, the best systems to run their code, or start an in-depth code analysis.
To submit a code, a user/developer must supply an ‘example’ benchmark of their code, i.e., a job configuration which is:
- Small such that the analyst will not wait for hours or days.
- Indicative of a typical workload.
- Has to come along with all the descriptions such that RTPs can work on the code independently.
A submission has to enable a small team of RPTs or individuals to take a submitted code and make it run through a standardised black-box assessment. For this, code submitters must provide enough information, so the RTP team can take the code and start working without a hassle. This means, the submitting group has to provide some “at least” information, which could comprise (but is not limited to):
- Readme files with build instructions
- Self-contained archive or download instructions
- Containers
- Invocation scripts
- One single batch file that builds and executes everything
- Information on resource use
It is the job of this workpackage to identify what kind of (reproducibility) information we expect external codes to provide. For this, we have to study similar initiatives (cmp SSI, POP, but also the SC reproducibility badge) and distill it into a list of checkboxes. Obviously, a project tackling this tasks might apply the guidelines immediately to some of their codes.