Skip to main content
Log in

An empirical evaluation of cross project priority prediction

  • Original Article
  • Published:
International Journal of System Assurance Engineering and Management Aims and scope Submit manuscript

Abstract

Prioritization of bugs decides the bug fix sequence. Incorrect prioritization of bugs results in delay of resolving the important bugs, which leads delay in release of the software. Prediction of bug priority needs historical data on which we can train the classifiers. However, such historical data is not always available in practice. In this situation, building classifiers based on cross project is the solution. In the available literature, we found very few papers for bug priority prediction and none of them dealt in cross project context. In this paper, we have evaluated the performance of different machine learning techniques namely Support Vector Machine, Naive Bayes (NB), K-Nearest Neighbors and Neural Network in predicting the priority of the newly coming reports in intra and cross project context. To evaluate the performance of these machine learning techniques for priority prediction in cross project context, we have considered three scenarios: (i) 10 fold cross-validation for intra project (ii) cross project validation for inter projects and (iii) inter project cross validation with different combination of training datasets. We performed experiments for each scenario on five datasets. Results from these experiments conclude that the accuracy performance for all machine learning techniques except NB is above 70, 72 and 73 % in respective scenarios. The experimental results also show that the combination of different project datasets for training candidates does not provide a significant improvement in performance measures.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Institutional subscriptions

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

References

  • Anvik J (2006) Automating bug report assignment. In: Proceedings of the 28th International Conference on Software Engineering, Shanghai, China, pp 937–940

  • Anvik J, Murphy GC (2011) Reducing the effort of bug report triage: Recommenders for development-oriented decisions. ACM Transact Softw Eng Methodol 20(3):10

    Google Scholar 

  • Anvik J, Hiew L, Murphy GC (2006) Who should fix this bug? In: Proceedings of the 28th International conference on Software Engineering, Shanghai, China, pp 361–370

  • Canfora G, Cerulo L (2006) Supporting change request assignment in open source development. In: Proceedings of the ACM Symposium on Applied Computing, Dijon, France, pp 1767–1772

  • Chaturvedi KK, Singh VB (2012) Determining bug severity using machine learning techniques. In: Proceedings of the International Conference on Software Engineering (CONSEG), Indore, India, pp 378–387

  • Denil M, Trappenberg T (2010) Overlap versus Imbalance. In: Proceedings of the 23rd Canadian Conference on Advances in Artificial Intelligence, Springer, Verlag Berlin, Heidelberg, pp 220–231

  • He Z, Shu F, Yang Y, Li M, Wang Q (2012) An investigation on the feasibility of cross-project defect prediction. In: Automated Software Engineering, pp 167–199

  • Kanwal J, Maqbool O (2010) Managing open bug repositories through bug report prioritization using SVMs. In: Proceedings of the International Conference on Open-Source Systems and Technologies, Lahore, Pakistan

  • Kanwal J, Maqbool O (2012) Bug prioritization to facilitate bug report triage. J Comput Sci Technol 27(2):397–412

    Article  Google Scholar 

  • Kim S, Whitehead J (2006) How long did it take to fix bugs? In: Proceedings of the International Workshop on Mining Software Repositories, Shanghai, China, pp 173–174

  • Lamkanfi A, Demeyer S, Gigery E, Goethals B (2010) Predicting the severity of a reported bug. In: Proceedings of the 7th Working Conference on Mining Software Repositories, Cape Town, South Africa, pp 1–10

  • Marks L, Zou YA, Hassan E (2011) Studying the fix-time for bugs in large open source projects. In: Proceedings of the 7th International Conference on Predictive Models in Software Engineering, Banff, Article No. 11

  • Menzies T, Marcus A (2008) Automated severity assessment of software defect reports. In: Proceedings of the International Conference on Software Maintenance, IEEE, pp 346–355

  • Mierswa I, Wurst M, Klinkenberg R, Scholz M, Euler T (2006) YALE: Rapid Prototyping for Complex Data Mining Tasks. In: Proceedings of the 12th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining (KDD-06). http://www.rapid-i.com

  • Porter M (2008) An algorithm for suffix stripping. Program 14(3):130–137

    Article  Google Scholar 

  • Sharma M, Bedi P, Chaturvedi KK, Singh VB (2012) Predicting the priority of a reported bug using machine learning techniques and cross project validation. In: Proceedings of the 12th International Conference on Intelligent Systems Design and Applications (ISDA) Kochi, India, pp 539–545

  • Sharma M, Kumari M, Singh VB (2013) Understanding the meaning of bug attributes and prediction models. In: Proceedings of the 5th IBM Collaborative Academia Research Exchange Workshop, I-CARE 2013, Article No. 15, ACM, New York, USA

  • Tamrawi A, Nguyen T, Al-Kofahi J, Nguyen TN (2011) Fuzzy set based automatic bug triaging. In: Proceedings of the 33rd International conference on Software Engineering (NIER Track), Miami, USA, pp 884–887

  • Turhan B, Menzies T, Bener AB, Stefano JD (2009) On the relative value of cross-company and within-company data for defect prediction. Empir Softw Eng. doi:10.1007/s10664-008-9103-7

    Google Scholar 

  • Weib C, Premraj R, Zimmermann T, Zeller A (2007) Predicting effort to fix software bugs. In: Proceedings of the Workshop on Software Reengineering, Bad Honnef, Germany

  • Yu L, Tsai W, Zhao W, Wu F (2010) Predicting defect priority based on neural networks. In: Proceedings of the 6th International Conference on Advanced Data Mining and Applications, Wuhan, China, pp 356–367

  • Zimmermann T, Nagappan N, Gall H (2009) Cross-project defect prediction: a large scale experiment on data vs. domain vs. process. In: Proceedings of the 7th Joint Meeting of the European Software Enginnering Conference and the ACM SIGSOFT Symposium on The Foundations of Software Engineering, pp 91–100

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Meera Sharma.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Sharma, M., Bedi, P. & Singh, V.B. An empirical evaluation of cross project priority prediction. Int J Syst Assur Eng Manag 5, 651–663 (2014). https://doi.org/10.1007/s13198-014-0219-4

Download citation

  • Received:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s13198-014-0219-4

Keywords

Navigation