Home > Research > Publications & Outputs > Researcher bias

Electronic data

  • Bias

    Rights statement: ©2014 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.

    Accepted author manuscript, 431 KB, PDF document

    Available under license: CC BY-NC: Creative Commons Attribution-NonCommercial 4.0 International License

Links

Text available via DOI:

View graph of relations

Researcher bias: The use of machine learning in software defect prediction

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Published

Standard

Researcher bias: The use of machine learning in software defect prediction. / Shepperd, Martin; Bowes, David; Hall, Tracy.
In: IEEE Transactions on Software Engineering, Vol. 40, No. 6, 6824804, 01.06.2014, p. 603-616.

Research output: Contribution to Journal/MagazineJournal articlepeer-review

Harvard

Shepperd, M, Bowes, D & Hall, T 2014, 'Researcher bias: The use of machine learning in software defect prediction', IEEE Transactions on Software Engineering, vol. 40, no. 6, 6824804, pp. 603-616. https://doi.org/10.1109/TSE.2014.2322358

APA

Shepperd, M., Bowes, D., & Hall, T. (2014). Researcher bias: The use of machine learning in software defect prediction. IEEE Transactions on Software Engineering, 40(6), 603-616. Article 6824804. https://doi.org/10.1109/TSE.2014.2322358

Vancouver

Shepperd M, Bowes D, Hall T. Researcher bias: The use of machine learning in software defect prediction. IEEE Transactions on Software Engineering. 2014 Jun 1;40(6):603-616. 6824804. doi: 10.1109/TSE.2014.2322358

Author

Shepperd, Martin ; Bowes, David ; Hall, Tracy. / Researcher bias : The use of machine learning in software defect prediction. In: IEEE Transactions on Software Engineering. 2014 ; Vol. 40, No. 6. pp. 603-616.

Bibtex

@article{7ec00a72e7b745d6a3a3e6dfc2829916,
title = "Researcher bias: The use of machine learning in software defect prediction",
abstract = "Background. The ability to predict defect-prone software components would be valuable. Consequently, there have been many empirical studies to evaluate the performance of different techniques endeavouring to accomplish this effectively. However no one technique dominates and so designing a reliable defect prediction model remains problematic. Objective. We seek to make sense of the many conflicting experimental results and understand which factors have the largest effect onpredictive performance. Method. We conduct a meta-analysis of all relevant, high quality primary studies of defect prediction to determine what factors influence predictive performance. This is based on 42 primary studies that satisfy our inclusion criteria that collectively report 600 sets of empirical prediction results. By reverse engineering a common response variable we build arandom effects ANOVA model to examine the relative contribution of four model building factors (classifier, data set, input metrics and researcher group) to model prediction performance. Results. Surprisingly we find that the choice of classifier has little impact upon performance (1.3 percent) and in contrast the major (31 percent) explanatory factor is the researcher group. It matters more who does the work than what is done. Conclusion. To overcome this high level of researcher bias, defect prediction researchers should (i) conduct blind analysis, (ii) improve reporting protocols and (iii) conduct more intergroup studies in order to alleviate expertise issues. Lastly, research is required to determine whether this bias is prevalent in other applications domains.",
keywords = "meta-analysis, researcher bias, Software defect prediction",
author = "Martin Shepperd and David Bowes and Tracy Hall",
note = "{\textcopyright}2014 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.",
year = "2014",
month = jun,
day = "1",
doi = "10.1109/TSE.2014.2322358",
language = "English",
volume = "40",
pages = "603--616",
journal = "IEEE Transactions on Software Engineering",
issn = "0098-5589",
publisher = "Institute of Electrical and Electronics Engineers Inc.",
number = "6",

}

RIS

TY - JOUR

T1 - Researcher bias

T2 - The use of machine learning in software defect prediction

AU - Shepperd, Martin

AU - Bowes, David

AU - Hall, Tracy

N1 - ©2014 IEEE. Personal use of this material is permitted. However, permission to reprint/republish this material for advertising or promotional purposes or for creating new collective works for resale or redistribution to servers or lists, or to reuse any copyrighted component of this work in other works must be obtained from the IEEE.

PY - 2014/6/1

Y1 - 2014/6/1

N2 - Background. The ability to predict defect-prone software components would be valuable. Consequently, there have been many empirical studies to evaluate the performance of different techniques endeavouring to accomplish this effectively. However no one technique dominates and so designing a reliable defect prediction model remains problematic. Objective. We seek to make sense of the many conflicting experimental results and understand which factors have the largest effect onpredictive performance. Method. We conduct a meta-analysis of all relevant, high quality primary studies of defect prediction to determine what factors influence predictive performance. This is based on 42 primary studies that satisfy our inclusion criteria that collectively report 600 sets of empirical prediction results. By reverse engineering a common response variable we build arandom effects ANOVA model to examine the relative contribution of four model building factors (classifier, data set, input metrics and researcher group) to model prediction performance. Results. Surprisingly we find that the choice of classifier has little impact upon performance (1.3 percent) and in contrast the major (31 percent) explanatory factor is the researcher group. It matters more who does the work than what is done. Conclusion. To overcome this high level of researcher bias, defect prediction researchers should (i) conduct blind analysis, (ii) improve reporting protocols and (iii) conduct more intergroup studies in order to alleviate expertise issues. Lastly, research is required to determine whether this bias is prevalent in other applications domains.

AB - Background. The ability to predict defect-prone software components would be valuable. Consequently, there have been many empirical studies to evaluate the performance of different techniques endeavouring to accomplish this effectively. However no one technique dominates and so designing a reliable defect prediction model remains problematic. Objective. We seek to make sense of the many conflicting experimental results and understand which factors have the largest effect onpredictive performance. Method. We conduct a meta-analysis of all relevant, high quality primary studies of defect prediction to determine what factors influence predictive performance. This is based on 42 primary studies that satisfy our inclusion criteria that collectively report 600 sets of empirical prediction results. By reverse engineering a common response variable we build arandom effects ANOVA model to examine the relative contribution of four model building factors (classifier, data set, input metrics and researcher group) to model prediction performance. Results. Surprisingly we find that the choice of classifier has little impact upon performance (1.3 percent) and in contrast the major (31 percent) explanatory factor is the researcher group. It matters more who does the work than what is done. Conclusion. To overcome this high level of researcher bias, defect prediction researchers should (i) conduct blind analysis, (ii) improve reporting protocols and (iii) conduct more intergroup studies in order to alleviate expertise issues. Lastly, research is required to determine whether this bias is prevalent in other applications domains.

KW - meta-analysis

KW - researcher bias

KW - Software defect prediction

U2 - 10.1109/TSE.2014.2322358

DO - 10.1109/TSE.2014.2322358

M3 - Journal article

AN - SCOPUS:84903176990

VL - 40

SP - 603

EP - 616

JO - IEEE Transactions on Software Engineering

JF - IEEE Transactions on Software Engineering

SN - 0098-5589

IS - 6

M1 - 6824804

ER -