Home > Research > Publications & Outputs > Reproducibility and replicability of software d...

Electronic data

  • ist2018FINAL

    Rights statement: This is the author’s version of a work that was accepted for publication in Information and Software Technology. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. A definitive version was subsequently published in Information and Software Technology, 99, 2018 DOI: 10.1016/j.infsof.2018.02.003

    Accepted author manuscript, 416 KB, PDF document

    Available under license: CC BY-NC-ND

Links

Text available via DOI:

View graph of relations

Reproducibility and replicability of software defect prediction studies

Research output: Contribution to Journal/MagazineReview articlepeer-review

Published
<mark>Journal publication date</mark>1/07/2018
<mark>Journal</mark>Information and Software Technology
Volume99
Number of pages16
Pages (from-to)148-163
Publication StatusPublished
Early online date10/02/18
<mark>Original language</mark>English

Abstract

Context: Replications are an important part of scientific disciplines. Replications test the credibility of original studies and can separate true results from those that are unreliable. Objective: In this paper we investigate the replication of defect prediction studies and identify the characteristics of replicated studies. We further assess how defect prediction replications are performed and the consistency of replication findings. Method: Our analysis is based on tracking the replication of 208 defect prediction studies identified by a highly cited Systematic Literature Review (SLR) [1]. We identify how often each of these 208 studies has been replicated and determine the type of replication carried out. We identify quality, citation counts, publication venue, impact factor, and data availability from all 208 SLR defect prediction papers to see if any of these factors are associated with the frequency with which they are replicated. Results: Only 13 (6%) of the 208 studies are replicated. Replication seems related to original papers appearing in the Transactions of Software Engineering (TSE) journal. The number of citations an original paper had was also an indicator of replications. In addition, studies conducted using closed source data seems to have more replications than those based on open source data. Where a paper has been replicated, 11 (38%) out of 29 studies revealed different results to the original study. Conclusion: Very few defect prediction studies are replicated. The lack of replication means that it remains unclear how reliable defect prediction is. We provide practical steps for improving the state of replication.

Bibliographic note

This is the author’s version of a work that was accepted for publication in Information and Software Technology. Changes resulting from the publishing process, such as peer review, editing, corrections, structural formatting, and other quality control mechanisms may not be reflected in this document. Changes may have been made to this work since it was submitted for publication. A definitive version was subsequently published in Information and Software Technology, 99, 2018 DOI: 10.1016/j.infsof.2018.02.003