Keywords
Software, Best Practices, Metrics, Evaluation, Impact, ELIXIR, Sustainability
This article is included in the ELIXIR gateway.
This article is included in the Bioinformatics Education and Training Collection collection.
This article is included in the EMBL-EBI collection.
Software, Best Practices, Metrics, Evaluation, Impact, ELIXIR, Sustainability
Compliance with and promotion of good development practice is a powerful mechanism for promoting software sustainability. Using metrics to judge good practice can enhance research software maintainability and helps establish a baseline of quality, reusability and reproducibility. Software development metrics, however, are only useful if it is clear what they measure. This could be a) the application of agreed good practice in a piece of software or software team, or b) how sustainable the software is in the long term. There have been previous attempts to assess good practices for scientific computing1 but they did not specifically tackle the question how to measure them during software development. As part of a collaboration between the ELIXIR pan-European research infrastructure for life science data and the Software Sustainability Institute, a working group met at Schiphol airport (Amsterdam) on December 14–15th 2015 to a) define and select the metrics that reflect the application of good practices, b) discuss the collection of these metrics and c) establish how the metrics could be implemented to ensure their wide adoption. In this article we report the outcomes of this workshop. We believe this effort is set apart from previous initiatives because of its ‘bottom up’ approach to ensure community adoption and therefore it should have realistic chances of implementation. We benefit from the fact that participating members of both groups have long established track records in life science research software development. This is the first release of our agreed software development good practices and expect that new revisions could evolve from it in future iterations. It is outside of the scope of this manuscript to delve into the issues that these metrics might raise in terms of performance comparison between different software.
In a workshop 12 experts from across Europe met to discuss good software development practices for life sciences. At the meeting, the group was divided randomly into two equally large subgroups to facilitate discussion, each subgroup spending a set time discussing potential metrics, their impact and applicability. The experts in each subgroup did not impose any restriction on which metrics to propose, but rather aimed to be as inclusive as possible, as long as each suggested metric had potential for impact. After the discussion, each group summarised the results and subsequently we merged the resulting metrics together into a list of 17 topics.
Next, the two groups worked on prioritising the identified metrics according to two criteria: 1) Importance and 2) Implementability. Importance is a measure of the impact that a particular practice can have in making software more sustainable. A metric is considered highly implementable if it is easy to generate. For each identified metric, importance and implementability were ranked by all members of the working group on a scale from 1 to 5, 5 being highest importance or easiest implementation. An average score was calculated and the resulting list was sorted from highest to lowest scoring metrics. Here we discuss and evaluate a final list of the top ten prioritised metrics.
We identified a set of 17 topics that are critical for software development good practice (Box 1). It was evident that these include measurements of different styles: measurements can be self-reported, automatically produced or externally audited. The type of metric is also important to consider here: there are metrics of qualitative and quantitative nature. Qualitative metrics correspond to a binary classification description, while quantitative metrics tend to be more amenable to integration and presentation as statistics. Metrics interpretation may pose challenges of its own kind, particularly related to the subjective nature of the importance of metrics and the different perceptions of value according to the context in which they are used.
Each of these topics have quantitative and qualitative metrics that may help track the adoption of good practice and monitor compliance with the guidelines in life sciences.
1. Version control:
2. Code reviews:
3. Automated testing:
4. Not reinventing the wheel:
5. Discoverability:
6. Reusability of source code:
7. Reusability of software:
a. Number of citations on the paper
b. Having basic description of features in structured ELIXIR format (EDAM ontology4) - in ELIXIR Tools and Data Services Registry?
8. Licensing:
9. Issue tracking/bug tracking:
10. Support processes:
11. Compliance with community standards:
12. Buildable code:
13. Open development:
14. Making data available:
15. Documentation:
16. Simplicity:
17. Dependency management:
We used the 43 metrics contained in the 17 identified topics as a basis for further prioritisation as described in the Methods section. Prioritisation of metrics was achieved by all participants scoring them according to their perception of importance and implementability. An average score was calculated and a sum of average importance and average implementability to rank the list (Table 1). We introduced also a manual evaluation for each of the proposed ranked metrics, which reflected the consensus of the final prioritisation, given initial difference of opinions when reviewing the average scores. In Table 1, we summarise the top 10 suggested metrics.
Each identified metric was scored according to importance (for sustainability) and implementability. Importance scores ranged from 1 (little) to 5 (very much) and implementability from 1 (difficult) to 5 (easy). Average values are shown for both importance (a) and implementability (b). A priority score (c) is calculated as the sum of the averages provided by (a) and (b). (c) is further discussed and the final Manual Priority Evaluation (d) is agreed, reflecting the final prioritisation judgement decided by the Working Group.
As a use case, we base the application of these metrics within the context of code development in ELIXIR. We define each of the 10 prioritised metrics in Table 1 and, where necessary, describe and explain the motivation for a metric and how to measure it. We consider that these definitions are applicable to a wider range of software development communities in life sciences.
1. Is controlled versioning used?
○ Description: Is it clearly indicated, can it be easily found?
○ Motivation: Version control systems provide an environment for safe and transparent software development.
○ How to measure: Put information about a version control tool to the ELIXIR Tools and Data Services Registry (which system, when it was installed, …)
2. Is the software discoverable?
○ Description: Is it easy to find the software based on its functionality (without knowing its exact name)?
○ Motivation: It is important to be discoverable so other potential contributions are encouraged and more people use the software.
○ How to measure: The ELIXIR community should be motivated and guided to provide this information into the ELIXIR Tools and Data Services Registry. If not, a list of other catalogues should be defined (maximum 5–10 other sources, e.g. BioSharing, field-specific catalogues, etc.). If the tool cannot be found there, the discoverability should be evaluated as 0.
3. Is an automated build system used?
○ Description: Are the builds of the software performed by some automated system?
○ Motivation: If the automated system for builds is applied, can the users rebuild the software easily, which markedly increases its usability?
○ How to measure: This information should be again included into the ELIXIR Tools and Data Services Registry2. Ideally, a link to the installation document should also be provided. How many commands are necessary for building of the software? (Optimally, just one command should be performed.)
4. Are test data available?
○ Description: Are data for testing of the software easily available for users?
○ Motivation: Without test data, it may be difficult to try the functionality of the software and assess correct functioning of an installation.
○ How to measure: The test data should be linked to from the web page describing the software or in the supplementary material of its associated publication. A link to the data should be included in ELIXIR Service Registry.
5. Does software contain parts that reimplement existing technology?
○ Description: Are common components/algorithms covered by libraries or reimplemented?
○ Motivation: A (naïve) reimplementation can cause unnecessary errors or decrease the effectiveness.
○ How to measure: Percentage of code written from scratch and/or number of used libraries. Additionally, descriptions of why a library with similar functionality was not used and responses to suggestions from community.
6. Does the software support open community standards and what is its level of compliance?
○ Description: Evaluation of software compliance with open/community standards
○ Motivation: This is needed, for example, where data input/output, networking and general interoperability are concerned. However, it is also non-trivial to implement and measure in terms of the overall software quality.
○ How to measure: A base metric would be: “does the software make use of open standards (yes/no), if so which ones (listing)?” In addition, more qualitative information such as “which versions of the standard does the software support?”, “Is it compatible with the latest specification?”, and “Can it be used to provide a more general level of support?” Another fundamental aspect to consider is whether the standard provide its own compliance metric (e.g., a test suite) and what the software’s level of compliance is. An example of such a compliance test suite is provided by the Systems Biology Markup Language (SBML,5).
7. Are code reviews performed?
○ Description: Whether new code is inspected by someone else before it becomes part of the code base.
○ Motivation: Code reviews increase quality of the code both because it is written with more care and because the second pair of eyes will more readily catch false assumptions or errors.
○ How to measure: Activity in code review process (comments to updated lines, etc.)
8. Is automated testing performed?
○ Description: Is some system for automated testing implemented?
○ Motivation: Automatic testing decreases occurrence of bugs.
○ How to measure: Information about the testing methodology should be present in the software documentation. In parallel, developers can be motivated to add this information to ELIXIR Service Registry.
9. Is the code documented?
○ Description: Does the code contain comments describing its main elements?
○ Motivation: Code comments increase the readability of the code and also indirectly motivate the programmer to write a cleaner code. However, commenting can present the problem of not being updated as code changes. This means that code comments may rot and become misleading/inaccurate. Often comments can be made redundant by better names of variables and methods. An exception is example code where explaining what each line does with a comment is useful.
○ How to measure: Determine the percentage of text from the source code that corresponds to comments.
10. How high is the code complexity?
We present an initial set of 10 good practices that could help make software for the life sciences more sustainable. From our discussions, it was clear that a community-wide adoption of standards is needed in terms of how measurement of metrics are collected and shared. We operate under the assumption that all software developed should be open source from the beginning of development, which means that the collection of statistics for good practice compliance should not violate any of the licensing or privacy issues associated to closed code.
These ‘Top 10 Good Practices’ should be considered as an initial view of what the community considers important with a description of their feasibility for implementation within the life sciences. Among our top suggested topics there is a remarkable coincidence on the need for versioning. The ways on how to collect metrics regarding versioning systems vary: if using GitHub, a number of statistics are readily available that allow their easy collection for benchmarking. We do not, however, want to prescribe which versioning systems should be adopted. There are many ways in which this metric can be measured, a sample of which we offer. The metrics we propose can be both qualitative and quantitative. Although quantitative metrics are easier to track, it is also important to capture qualitative characteristics such as existence of automated testing or compliance with community standards.
This article is a first attempt to crystallise the conclusions from the work that the group of experts gathered under the auspices of ELIXIR and the Software Sustainability Institute. It is thus not intended to be a final declaration of what the ELIXIR community thinks the metrics, implementation and feasibility for measuring good practices for software development should be. This document is an initial response from the working group established to assess the problem of evaluating metrics for software development good practices. We expect it to be modified in future versions as more experts join this group and new challenges emerge with evolving technologies and life science software needs.
All authors participated in the discussions, selection and prioritisation of metrics. We believe all authors contributed equally to this work. All authors contributed to the writing of this article. All authors read and approved the submitted manuscript.
MC and HA are strategically core funded by UK’s BBSRC. BGO is funded by the BE-BASIC grant F08.005. NCH was supported by EPSRC, BBSRC and ESRC Grant EP/N006410/1 for the UK Software Sustainability Institute. The work was part of the ELIXIR-EXCELERATE project, funded by the European Commission within the Research Infrastructures programme of Horizon 2020, grant agreement number 676559.
The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.
We are grateful to Susanna Repo (ELIXIR-Hub) and Chris Ponting (ELIXIR-UK Head of Node) for comments and feedback.
Views | Downloads | |
---|---|---|
F1000Research | - | - |
PubMed Central
Data from PMC are received and updated monthly.
|
- | - |
References
1. Giannoulatou E, Park SH, Humphreys DT, Ho JW: Verification and validation of bioinformatics software without a gold standard: a case study of BWA and Bowtie.BMC Bioinformatics. 2014; 15 Suppl 16: S15 PubMed Abstract | Publisher Full TextCompeting Interests: No competing interests were disclosed.
Competing Interests: No competing interests were disclosed.
Alongside their report, reviewers assign a status to the article:
Invited Reviewers | ||
---|---|---|
1 | 2 | |
Version 1 16 Aug 16 |
read | read |
Provide sufficient details of any financial or non-financial competing interests to enable users to assess whether your comments might lead a reasonable person to question your impartiality. Consider the following examples, but note that this is not an exhaustive list:
Sign up for content alerts and receive a weekly or monthly email with all newly published articles
Already registered? Sign in
The email address should be the one you originally registered with F1000.
You registered with F1000 via Google, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Google account password, please click here.
You registered with F1000 via Facebook, so we cannot reset your password.
To sign in, please click here.
If you still need help with your Facebook account password, please click here.
If your email address is registered with us, we will email you instructions to reset your password.
If you think you should have received this email but it has not arrived, please check your spam filters and/or contact for further assistance.
Thank you authors for discussing the main factors for developing and maintaining reliable, reusable, and durable software, as these are the main prerequisites for ... Continue reading A great and timely article!
Thank you authors for discussing the main factors for developing and maintaining reliable, reusable, and durable software, as these are the main prerequisites for further good practices towards scientific software's qualities such as efficiency, accessibility, or interoperability.
One crucial factor appears implicitly throughout the article, but isn't phrased explicitly: the importance of having a public source code repository, with compilable and debuggable code, and installable binaries. And perhaps also a good support for building a community of contributors (Budd et al. 2015), transparent and participatory, with social media, mailing lists, events, public discussions ...
I'm personally also a bit surprised that especially licensing hasn't been voted higher. And issue tracking neither. I could also imagine more explicit references to free software and open source, to FSF and OSI, and maybe to Stallman 1986, Perens 1999, or also 1997.
Altogether, I'm very glad to see an article like this published!
Matúš Kalaš
References:
Budd, A. et al. (2015). A quick guide for building a successful bioinformatics community. PLoS Comput. Biol., 11(2), e1003972.
Perens, B. (1997). Debian’s ”Social Contract” with the Free Software Community. debian-announce@lists.debian.org, (msg00017). Re-published as Debian Social Contract, Version 1.0.
Perens, B. (1999). The Open Source Definition. In Open Sources: Voices from the Open Source Revolution. O’Reilly.
Stallman, R. M. (1986). What is the Free Software Foundation? Gnu’s Bulletin, 1(1), 8–9.
Thank you authors for discussing the main factors for developing and maintaining reliable, reusable, and durable software, as these are the main prerequisites for further good practices towards scientific software's qualities such as efficiency, accessibility, or interoperability.
One crucial factor appears implicitly throughout the article, but isn't phrased explicitly: the importance of having a public source code repository, with compilable and debuggable code, and installable binaries. And perhaps also a good support for building a community of contributors (Budd et al. 2015), transparent and participatory, with social media, mailing lists, events, public discussions ...
I'm personally also a bit surprised that especially licensing hasn't been voted higher. And issue tracking neither. I could also imagine more explicit references to free software and open source, to FSF and OSI, and maybe to Stallman 1986, Perens 1999, or also 1997.
Altogether, I'm very glad to see an article like this published!
Matúš Kalaš
References:
Budd, A. et al. (2015). A quick guide for building a successful bioinformatics community. PLoS Comput. Biol., 11(2), e1003972.
Perens, B. (1997). Debian’s ”Social Contract” with the Free Software Community. debian-announce@lists.debian.org, (msg00017). Re-published as Debian Social Contract, Version 1.0.
Perens, B. (1999). The Open Source Definition. In Open Sources: Voices from the Open Source Revolution. O’Reilly.
Stallman, R. M. (1986). What is the Free Software Foundation? Gnu’s Bulletin, 1(1), 8–9.