Home > Research > Publications & Outputs > Further thoughts on precision

Links

Text available via DOI:

View graph of relations

Further thoughts on precision

Research output: Contribution in Book/Report/Proceedings - With ISBN/ISSNConference contribution/Paperpeer-review

Published
Close
Publication date2011
Host publication15th Annual Conference on Evaluation & Assessment in Software Engineering (EASE 2011)
PublisherIEEE
Pages129-133
Number of pages5
ISBN (print)9781849195096
<mark>Original language</mark>English

Abstract

Background: There has been much discussion amongst automated software defect prediction researchers regarding use of the precision and false positive rate classifier performance metrics. Aim: To demonstrate and explain why failing to report precision when using data with highly imbalanced class distributions may provide an overly optimistic view of classifier performance. Method: Well documented examples of how dependent class distribution affects the suitability of performance measures. Conclusions: When using data where the minority class represents less than around 5 to 10 percent of data points in total, failing to report precision may be a critical mistake. Furthermore, deriving the precision values omitted from studies can reveal valuable insight into true classifier performance.