Abstract
Defect prediction models can help Software Quality Assurance (SQA) teams understand their past pitfalls that lead to defective modules. However, the conclusions that are derived from defect prediction models without mitigating redundant metrics issues may be misleading. In this paper, we set out to investigate if redundant metrics issues are affecting defect prediction studies, and its degree and causes of redundancy. Through a case study of 101 publicly-available defect datasets of systems that span both proprietary and open source domains, we observe that (1) 10%-67% of metrics of the studied defect datasets are redundant, and (2) the redundancy of metrics has to do with the aggregation functions of metrics. These findings suggest that researchers should be aware of redundant metrics prior to constructing a defect prediction model in order to maximize internal validity of their studies.
Original language | English |
---|---|
Title of host publication | Proceedings - 2016 IEEE 27th International Symposium on Software Reliability Engineering Workshops |
Subtitle of host publication | ISSREW 2016 |
Editors | Jeremy Bradbury |
Place of Publication | Piscataway NJ USA |
Publisher | IEEE, Institute of Electrical and Electronics Engineers |
Pages | 51-52 |
Number of pages | 2 |
ISBN (Electronic) | 9781509036011 |
DOIs | |
Publication status | Published - 2016 |
Externally published | Yes |
Event | International Symposium on Software Reliability Engineering 2016 - Ottawa, Canada Duration: 23 Oct 2016 → 27 Oct 2016 Conference number: 27th http://2016.issre.net/ |
Conference
Conference | International Symposium on Software Reliability Engineering 2016 |
---|---|
Abbreviated title | ISSRE 2016 |
Country/Territory | Canada |
City | Ottawa |
Period | 23/10/16 → 27/10/16 |
Internet address |
Keywords
- Defect prediction models
- Redundant metrics
- Software quality assurance