Bottlenecks in Software Defect Prediction Implementation in Industrial Projects

Open access


Case studies focused on software defect prediction in real, industrial software development projects are extremely rare. We report on dedicated R&D project established in cooperation between Wroclaw University of Technology and one of the leading automotive software development companies to research possibilities of introduction of software defect prediction using an open source, extensible software measurement and defect prediction framework called DePress (Defect Prediction in Software Systems) the authors are involved in. In the first stage of the R&D project, we verified what kind of problems can be encountered. This work summarizes results of that phase.

[1] Catal, C. and Diri, B. (2009). A systematic review of software fault prediction studies. Expert Systems with Application, 36:7346-7354.

[2] Fenton, N. and Neil, M. (1999). A critique of software defect prediction models. IEEE Transactions on Software Engineering, 25:675-689.

[3] Hall, T., Beecham, S., Bowes, D., D., G., and Counsell, S. (2012). A systematic review of fault prediction performance in software engineering. IEEE Transactions on Software Engineering, 38:1276-1304.

[4] Jureczko, M. and Madeyski, L. (2010). Towards identifying software project clusters with regard to defect prediction. In Proceedings of the 6th International Conference on Predictive Models in Software Engineering, PROMISE '10, pages 9:1-9:10, New York, NY, USA. ACM.

[5] Khoshgoftaar, T., Allen, E., Hudepohl, J., and Aud, S. (1997). Application of neural networks to software quality modelling of a very large telecommunications system. IEEE Transactions on Neural Networks, 8:902-909.

[6] Khoshgoftaar, T. and Seliya, N. (2004). Comparative assessment of software quality classification techniques: An empirical case study. Empirical Software Engineering, 9:229-257.

[7] Khoshgoftaar, T. and Seliya, N. (2005). Assessment of a new three-group software quality classification technique: An empirical case study. Empirical Software Engineering, 10:183-218.

[8] Klas, M., Nakao, H., Elberzhager, F., and Munch, J. (2008). Predicting defect content and quality assurance effectiveness by combining expert judgment and defect data-a case study. Proceedings of the 19th International Symposium on Software Reliability Engineering, pages 17-26.

[9] Li, P., Herbsleb, J., Shaw, M., and Robinson, B. (2006). Experiences and results from initiating field defect prediction and product test prioritization efforts at abb inc. Proceedings of the 28th International Conference on Software Engineering, pages 413-422.

[10] Madeyski, L. and Jureczko, M. (2014). Which Process Metrics Can Significantly Improve Defect Prediction Models? An Empirical Study. Software Quality Journal.

[11] Madeyski, L. and Majchrzak, M. (2012). ImpressiveCode DePress (Defect Prediction for software systems) Extensible Framework. Available as an open source project from GitHub:

[12] Madeyski, L. and Majchrzak, M. (2014). Software Measurement and Defect Prediction with DePress Extensible Framework. Foundations and Computing and Decision Sciences (accepted).

[13] Ostrand, T. and Weyuker, E. (2002). The distribution of faults in a large industrial software system. SIGSOFT Software Engineering Notes, 27:55-64.

[14] Ostrand, T., Weyuker, E., and Bell, R. (2005). Predicting the location and number of faults in large software systems. IEEE Transactions on Software Engineering, 31:340-355.

[15] Ostrand, T., Weyuker, E., and Bell, R. (2010). Programmer-based fault prediction. Proceedings of the Sixth International Conference on Predictive Models in Software Engineering, pages 1-10.

[16] Rudiger, L., Lundberg, J., and Lowe, W. (2008). Comparing software metrics tools. Proceedings of The 2008 International Symposium on Software Testing and Analysis, pages 131-142.

[17] Sliwierski, J., Zimmermann, T., and Zeller, A. (2005). When do changes induce fixes? Proceedings of The 2005 International Workshop on Mining Software Repositories.

[18] Succi, G., Pedrycz, W., Stefanovic, M., and Miller, J. (2003). Practical assessment of the models for identification of defect-prone classes in object-oriented commercial systems using design metrics. Journal of Systems and Software, 65:1-12.

[19] Tosun, A., Bener, B., Turhan, B., and Menzies, T. (2010). Practical considerations in deploying statistical methods for defect prediction: A case study within the turkish telecommunications industry. Information and Software Technology, 52:1242-1257.

[20] Tosun, A., Turhan, B., and Bener, A. (2009). Practical considerations in deploying ai for defect prediction: A case study within the turkish telecommunication industry. Proceedings of the Fifth International Conference on Predictor Models in Software Engineering, page 11.

[21] Turhan, B., Kocak, G., and Bener, A. (2009a). Data mining source code for locating software bugs: A case study in telecommunication industry. Expert Systems with Applications, 36:9986-9990.

[22] Turhan, B., Menzies, T., Bener, A., and Di Stefano, J. (2009b). On the relative value of cross-company and within-company data for defect prediction. Empirical Software Engineering, 14:540-578.

[23] Wong, W., Horgan, J., Syring, M., Zage, W., and Zage, D. (2000). Applying design metrics to predict fault-proneness: A case study on a large-scale software system. Software: Practice and Experience, 30:1587-1608.

Foundations of Computing and Decision Sciences

The Journal of Poznan University of Technology

Journal Information

CiteScore 2017: 0.82

SCImago Journal Rank (SJR) 2017: 0.212
Source Normalized Impact per Paper (SNIP) 2017: 0.523

Mathematical Citation Quotient (MCQ) 2017: 0.02

Cited By


All Time Past Year Past 30 Days
Abstract Views 0 0 0
Full Text Views 160 160 15
PDF Downloads 43 43 3