Abstract
Software testing (ST) has been considered as one of the most important and critical activities of the software development life cycle (SDLC) since it influences directly on quality. When a software project is planned, it is common practice to predict the corresponding ST effort (STEP) as a percentage of predicted SDLC effort. However, the effort range for ST has been reported between 10 and 60% of the predicted SDLC effort. This wide range on STEP causes uncertainty in software managers due to STEP is used for allocating resources to teams exclusively for testing activities, and for budgeting and bidding the projects. In spite of this concern, hundreds of studies have been published since 1981 about SDLC effort prediction models, and only thirty-one STEP studies published in the last two decades were identified (just two of them based their conclusions on statistical significance). The contribution of the present study is to investigate the application for STEP of five machine learning (ML) models reported as the most accurate ones when applied to SDLC effort prediction. The models were trained and tested with data sets of projects selected from an international public repository of software projects. The selection for projects was based on their data quality rating, type of development, development platform, programming language generation, sizing method, and resource level of projects. Results based on statistical significance allow suggesting the application of specific ML models to software projects by type of development, and developed on a determined platform and programming language generation.


Similar content being viewed by others
References
Abhilasha, Sharma, A. (2013). Test effort estimation in regression testing. In: IEEE International Conference in MOOC, Innovation and Technology in Education (MITE), Jaipur, India, pp. 343-348. https://doi.org/10.1109/MITE.2013.6756364.
Abhishek, C., Kumar, V. P., Vitta, H., & Srivastava, P. R. (2010). Test effort estimation using neural network. Journal Software Engineering & Applications, 3, 331–340. https://doi.org/10.4236/jsea.2010.34038.
Ali, A., Gravino, C. (2019). A systematic literature review of software effort prediction using machine learning methods, Journal of Software: Evolution and Process, Wiley, 31(10), e2211. https://doi.org/10.1002/smr.2211.
Almeida, É.R.C., Abreu, B.T., Moraes, R. (2009). An alternative approach to test effort estimation based on use cases. In: IEEE International Conference on Software Testing Verification and Validation, pp. 279–288. https://doi.org/10.1109/ICST.2009.31.
Aloka, S., Singh, P., Rakshit, G., Srivastava, P.R. (2011). Test effort estimation-particle swarm optimization based approach, In: Communications in Computer and Information Science, pp. 463–474. https://doi.org/10.1007/978-3-642-22606-9_46.
Aranha, E., Borba, P. (2007). An estimation model for test execution effort. In: First International Symposium on Empirical Software Engineering and Measurement (ESEM), pp. 1–9. https://doi.org/10.1109/ESEM.2007.73.
Aranha, E., & Borba, P. (2015). Estimating manual test execution effort and capacity based on execution points. International Journal of Computers and Applications, Taylor & Francis, 31(3), 167–172. https://doi.org/10.1080/1206212X.2009.11441938.
Badri, M., & Toure, F. (2012). Empirical Analysis of object-oriented design metrics for predicting unit testing effort of classes. Journal of Software Engineering and Applications, 5, 513–526. https://doi.org/10.4236/jsea.2012.57060.
Badri, M., Toure, F., Lamontagne, L. (2015). Predicting unit testing effort levels of classes: An exploratory study based on multinomial logistic regression modeling. In: International Conference on Soft Computing and Software Engineering (SCSE), pp. 529 – 538. https://doi.org/10.1016/j.procs.2015.08.528.
Bardsiri, V. K., Jawawi, D. N. A., & Khatibi, E. (2014). Towards improvement of analogy-based software development effort estimation: A review. International Journal of Software Engineering and Knowledge Engineering (IJSEKE), 24(7), 1065–1089. https://doi.org/10.1142/S0218194014500351.
Bareja, K., Singhal, A. (2015). A review of estimation techniques to reduce testing efforts in software development. In: Fifth International Conference on Advanced Computing & Communication Technologies, pp. 541–546. https://doi.org/10.1109/ACCT.2015.110.
Bhattacharya, P., Srivastava, P.R., Prasad, B. (2012). Software test effort estimation using particle swarm optimization. In: International Conference on Information Systems Design and Intelligent Applications, Advances in Intelligent and Soft Computing, pp. 827–835. https://doi.org/10.1007/978-3-642-27443-5_95.
Bock F, German R, Siegl, S. (2016). Mathematical test effort estimation for dependability assessment of sensor-based driver assistance systems. In: 42th Euromicro Conference on Software Engineering and Advanced Applications, pp. 222–226. https://doi.org/10.1109/SEAA.2016.49.
Bock F, Siegl S, German, R. (2017). Analytical test effort estimation for multisensor driver assistance systems. In: IEEE 43rd Euromicro Conference on Software Engineering and Advanced Applications, pp. 239–246. https://doi.org/10.1109/SEAA.2017.49.
Boehm, B.W. (2017). Software cost estimation meets software diversity. In: IEEE/ACM 39th International Conference on Software Engineering Companion (ICSE-C), pp. 495–496. https://doi.org/10.1109/ICSE-C.2017.159.
Boser, B.E., Guyon, I.M., Vapnik, V.N. (1992). A training algorithm for optimal margin classifiers, In: 5th Annual ACM Workshop on Computational Learning Theory, pp. 144–152. https://doi.org/10.1145/130385.130401.
Bourque, P., Fairley, R. (2014). Guide to the software engineering body of knowledge, SWEBOK V3.0, IEEE Computer Society
Calzolari, F., Tonella, P., & Antonio, G. (2001). Maintenance and testing effort modeled by linear and nonlinear dynamic systems. Information and Software Technology, Elsevier, 43(8), 477–486. https://doi.org/10.1016/S0950-5849(01)00156-2.
Carbonera, C. E., Farias, K., & Bischoff, V. (2020). Software development effort estimation: A systematic mapping study. IET Software, 14(4), 328–344. https://doi.org/10.1049/iet-sen.2018.5334.
Chang, C. C., & Lin, C. J. (2011). LIBSVM: A library for support vector machines. ACM Transactions on Intelligent Systems and Technology, 2(3), 1–27. https://doi.org/10.1145/1961189.1961199.
Cortes, C., & Vapnik, V. N. (1995). Support-vector networks. Machine Learning., 20, 273–297. https://doi.org/10.1023/A:1022627411411.
Dawson, C. W. (1998). An artificial neural network approach to software testing effort estimation. WIT Transactions on Information and Communication Technologies, 20, 1–11. https://doi.org/10.2495/AI980361.
Ege-Adalı, O., Alpay, N., Gürel, Z. , Tahir, T., Gencel, C. (2017). Software test effort estimation, state of the art in Turkish software industry. In: IEEE 43rd Euromicro Conference on Software Engineering and Advanced Applications, pp. 412–420. https://doi.org/10.1109/SEAA.2017.72.
Felipe, N. F., Cavalcanti, R. P., Maia, E. H. B., Amaral, W. P., Farnese, A. C., Tavares, L. D., et al. (2014). A comparative study of three test effort estimation methods. Revista Cubana de Ciencias Informáticas, 8, 1–13.
Fernández-Diego, M., & González-Ladrón-de-Guevara, F. (2014). Potential and limitations of the ISBSG dataset in enhancing software engineering research: A mapping review. Information and Software Technology, Elsevier., 56(6), 527–544. https://doi.org/10.1016/j.infsof.2014.01.003.
Ferrucci, F., Gravino, C., Sarro, F. (2014). Exploiting prior-phase effort data to estimate the effort for the subsequent phases: A further assessment. In: 10th International Conference on Predictive Models in Software Engineering (PROMISE), pp. 42–51. https://doi.org/10.1145/2639490.2639509.
Finschi, I. (1996). An Implementation of The Levenberg-Marquardt Algorithm, Eidgenössische Technische Hochschule Zürich.
Fox, J. P. (2010). Bayesian item response modeling. Theory and Applications, Statistics for Social and Behavioral Sciences, Springer. https://doi.org/10.1007/978-1-4419-0742-4.
García-Floriano, A., Ferreira-Santiago, Á., Camacho-Nieto, O., & Yáñez-Márquez, C. (2019). A machine learning approach to medical image classification: Detecting age-related macular degeneration in fundus images. Computers & Electrical Engineering, Elsevier, 75, 218–229. https://doi.org/10.1016/j.compeleceng.2017.11.008.
Garousi, V., & Mäntylä, M. V. (2016). A systematic literature review of literature reviews in software testing. Information and Software Technology, Elsevier, 80, 195–216. https://doi.org/10.1016/j.infsof.2016.09.002.
Gass, S., & Fu, M. C. (2013). Lagrange Multipliers. Encyclopedia of operations research and management science: Springer.
Gautam, S.S., Singh, V. (2018). The state-of-the-art in software development effort estimation. Journal of Software: Evolution and Process, Wiley, e1983. https://doi.org/10.1002/smr.1983.
Grover, M., Bhatia, P.K., Mittal, H. (2017). Estimating software test effort based on revised UCP model using fuzzy technique. In: Information and Communication Technology for Intelligent Systems (ICTIS), pp. 490–498. https://doi.org/10.1007/978-3-319-63673-3_59.
Gupta, A., Tripathi, A., Kuswaha, D.S. (2015). Use case based approach to analyze software change impact and its regression test effort estimation. In: Advanced Computer and Communication Engineering Technology (LNEE), pp. 1057–1067. https://doi.org/10.1007/978-3-319-07674-4_98.
Hassouna, A., & Tahvildari, L. (2010). An effort prediction framework for software defect correction. Information and Software Technology, Elsevier, 52, 197–209. https://doi.org/10.1016/j.infsof.2009.10.003.
Haykin, S. (2009). Neural networks and learning machines. Pearson: Third Edition.
Humphrey, W.S. (1995). A Discipline for Software Engineering, First Edition, Addison-Wesley
Idri, A., Amazal, F. A., & Abran, A. (2015). Analogy-based software development effort estimation: A systematic mapping and review. Information and Software Technology. Elsevier, 58, 206–230. https://doi.org/10.1016/j.infsof.2014.07.013.
IFPUG. (2020). The International Function Point Users Group, http://www.ifpug.org/ . Accessed June 15, 2020
ISBSG. (2018). Guidelines for use of the ISBSG data, Release 2018, International Software Benchmarking Standards Group.
ISO/IEC 24570:2018. (2018). Software engineering, NESMA functional size measurement method, Definitions and counting guidelines for the application of function point analysis.
ISO/IEC 29881:2010. (2010). Information technology, Systems and software engineering, FiSMA 1.1 functional size measurement method.
Jayakumar, K. R., & Abran, A. (2013). A survey of software test estimation techniques. Journal of Software Engineering and Applications, 6, 47–52. https://doi.org/10.4236/jsea.2013.610A006.
Ji, H., Huang, S., Wu, Y., Hui, Z., & Zheng, C. (2019). A new weighted naive Bayes method based on information diffusion for software defect prediction. Software Quality Journal, Springer, 27, 923–968. https://doi.org/10.1007/s11219-018-9436-4.
Jorgensen, M., & Shepperd, M. (2007). A systematic review of software development cost estimation studies. IEEE Transactions on Software Engineering, 33(1), 33–53. https://doi.org/10.1109/TSE.2007.256943.
Kantardzic, M. (2011). Data Mining. Concepts, Models, Methods, and Algorithms, Second Edition, Wiley
Kaur, A., & Kaur, K. (2019). Investigation on test effort estimation of mobile applications: Systematic literature review and survey. Information and Software Technology, Elsevier, 110, 56–77. https://doi.org/10.1016/j.infsof.2019.02.003.
Khurana, P., Tripathi, A., Kushwaha, D.S. (2012). Change Impact Analysis and its Regression Test Effort Estimation. In: 3rd IEEE International Advance Computing Conference (IACC), pp. 1420–1424. https://doi.org/10.1109/IAdCC.2013.6514435.
Kitchenham, B. A., Pickard, L. M., MacDonell, S. G., & Shepperd, M. J. (2001). What accuracy statistics really measure. IEE Proceedings Software, 148(3), 81–85. https://doi.org/10.1049/ip-sen:20010506.
Kitchenham, B.A., Mendes, E. (2009). Why comparative effort prediction studies may be invalid. In: 5th International Conference on Predictor Models in Software Engineering (PROMISE), pp. 1–5. https://doi.org/10.1145/1540438.1540444.
Koza, J. R. (1998). Genetic Programming: On the programming of computers by means of natural selection. Cambridge: The MIT Press.
Koza, J. R. (1994). Genetic programming as a means for programming computers by natural selection. Statistics and Computing, Springer, 4, 87–112. https://doi.org/10.1007/BF00175355.
Kushwaha, D. S., & Misra, A. K. (2008). Software test effort estimation. ACM SIGSOFT Software Engineering Notes, 33(3), 1–6. https://doi.org/10.1145/1360602.1361211.
Langdon, W.B., Poli, R. (2011). Foundations of genetic programming, Springer-Verlag
Li, J. J., Ulrich, A., Bai, X., & Bertolino, A. (2020). Advances in test automation for software with special focus on artificial intelligence and machine learning. Software Quality Journal, Springer, 28, 245–248. https://doi.org/10.1007/s11219-019-09472-3.
Li, Z., Jing, X.-Y., & Zhu, X. (2018). Progress on approaches to software defect prediction. IET Software, 12(3), 161–175. https://doi.org/10.1049/iet-sen.2017.0148.
Lin, C. T., & Huang, C. Y. (2008). Enhancing and measuring the predictive capabilities of testing-effort dependent software reliability models. The Journal of Systems and Software, Elsevier, 81, 1025–1038. https://doi.org/10.1016/j.jss.2007.10.002.
Llermeier, E.H. (2007). Case-Based Approximate Reasoning, Springer
Lyu, M.R. (1996). Handbook of Software Reliability Engineering. IEEE Computer Society Press.
Mahmood, Y., Kama, N., Azmi, A. (2020). A systematic review of studies on use case points and expert‐based estimation of software development effort. Journal of Software: Evolution and Process, Wiley. e2245. https://doi.org/10.1002/smr.2245.
Malhotra, R., & Khanna, M. (2018). Threats to validity in search-based predictive modelling for software engineering. IET Software, 12(4), 293–305. https://doi.org/10.1049/iet-sen.2018.5143.
Mensah, S., Keung, J., Bennin, K.E., Bosu, M.F. (2016). Multi-Objective Optimization for Software Testing Effort Estimation. In: The 28th International Conference on Software Engineering and Knowledge Engineering, (SEKE), pp. 527–530. https://doi.org/10.18293/SEKE2016-017.
Moore, D. S., McCabe, G. P., & Craig, B. A. (2009). Introduction to the Practice of Statistics (6th ed.). Freeman and Company: W. H.
Nageswaran, S. (2001). Test effort estimation using use case points. In: 14th International Internet & Software Quality, pp. 1–6.
Nguyen, V., Pham, V., Lam, V. (2013). qEstimation: A Process for Estimating Size and Effort of Software Testing. In: International Conference on Software and System Process, pp. 20–28. https://doi.org/10.1145/2486046.2486052.
Prasad, D.S.U.M., Chacko, S., Kanakadandi, S.S.P., Durbhaka, G.K. (2014). Automated Regression Test Suite Optimization based on Heuristics. In: 4th International Conference on Artificial Intelligence with Applications in Engineering and Technology, pp: 48–53. https://doi.org/10.1109/ICAIET.2014.18.
Punitha, K., Chitra, S. (2013). Software defect prediction using software metrics - A survey. In: International Conference on Information Communication and Embedded Systems (ICICES), Chennai, India. https://doi.org/10.1109/ICICES.2013.6508369.
Richter, M. M., & Weber, R. O. (2013). Case-Based Reasoning. Springer. https://doi.org/10.1007/978-3-642-40167-1.
Rokach, L., & Maimon, O. (2005). Decision trees. Data Mining and Knowledge Discovery Handbook: Springer, Boston. https://doi.org/10.1007/0-387-25465-X_9.
Savolainen, P., Ahonen, J. J., & Richardson, I. (2012). Software development project success and failure from the supplier’s perspective: A systematic literature review. International Journal of Project Management, Elsevier, 30, 458–469. https://doi.org/10.1016/j.ijproman.2011.07.002.
Sette, S., & Boullart, L. (2001). Genetic programming: Principles and applications. Engineering Applications of Artificial Intelligence, Elsevier, 14(6), 727–736. https://doi.org/10.1016/S0952-1976(02)00013-1.
Sharma, A., & Kushwaha, D. S. (2012). Applying requirement based complexity for the estimation of software development and testing effort. ACM SIGSOFT Software Engineering Notes, 37(1), 1–11. https://doi.org/10.1145/2088883.2088898.
Sharma, A., & Kushwaha, D. S. (2013). An empirical approach for early estimation of software testing effort using SRS document. CSI Transactions on ICT, 1(1), 51–66. https://doi.org/10.1007/s40012-012-0003-z.
Shepperd, M., & MacDonell, S. (2012). Evaluating prediction systems in software project estimation. Information and Software Technology, Elsevier, 54(8), 820–827. https://doi.org/10.1016/j.infsof.2011.12.008.
Silva, D.G., Abreu, B.T., Jino, M. (2009). A simple approach for estimation of execution effort of functional test cases, In: International Conference on Software Testing Verification and Validation, pp. 289–298. https://doi.org/10.1109/ICST.2009.47.
Silva, D.G., Jino, M., Abreu, B.T. (2010). Machine learning methods and asymmetric cost function to estimate execution effort of software testing. In: Third IEEE International Conference on Software Testing, Verification and Validation, pp. 275–284. https://doi.org/10.1109/ICST.2010.46.
Singh, Y., Kaur, A., Malhotra, R. (2008). Predicting testing effort using artificial neural network, In: World Congress on Engineering and Computer Science (WCECS), pp. 1–6.
Song, Q., Shepperd, M., Cartwright, M., & Mair, C. (2006). Software defect association mining and defect correction effort prediction. IEEE Transactions on Software Engineering, 32(2), 69–82. https://doi.org/10.1109/TSE.2006.1599417.
Srivastava, P.R., Bidwai, A., Khan, A., Rathore, K., Sharma, R., Yang, X.S. (2014). An empirical study of test effort estimation based on bat algorithm, In: International Journal of Bio-Inspired Computation, pp. 57–70. https://doi.org/10.1504/IJBIC.2014.059966.
Srivastava, P. R., Kumar, S., Singh, A. P., & Raghurama, G. (2011). Software testing effort: An assessment through fuzzy criteria approach. Journal of Uncertain Systems, 5(3), 183–201.
Srivastava, PR, Varshney, A., Nama, P. (2012). Software test effort estimation: a model based on cuckoo search. International Journal of Bio-Inspired Computation, pp. 278–285. https://doi.org/10.1504/IJBIC.2012.049888.
Stewart, B. (2002). Predicting project delivery rates using the Naive-Bayes classifier. Journal of Software Maintenance and Evolution: Research and Practice, Wiley, 14, 161–179. https://doi.org/10.1002/smr.250.
Tiwari, R., & Goel, N. (2013). Reuse: Reducing test effort. ACM SIGSOFT Software Engineering Notes, 38(2), 1–11. https://doi.org/10.1145/2413038.2413042.
Tripathy, P., Naik, K. (2014). Software Evolution and Maintenance: A Practitioner's Approach, Wiley
Vapnik, V.N. (1998). Statistical Learning Theory, Wiley-Interscience
Veenendaal, E.P.W.M., Dekkers, T. (1999). Test point analysis: a method for test estimation. In: 10th European Software Control and Metrics conference and the 2nd SCOPE conference on software product evaluation, pp. 47–59.
Villuendas-Rey, Y., Rey-Benguría, C. F., Ferreira-Santiago, Á., Camacho-Nieto, O., & Yáñez-Márquez, C. (2017). The naïve associative classifier (NAC): a novel, simple, transparent, and accurate classification model evaluated on financial data. Neurocomputing, Elsevier, 265, 105–115. https://doi.org/10.1016/j.neucom.2017.03.085.
Wen, J., Li, S., Lin, Z., Hu, Y., & Huang, C. (2012). Systematic literature review of machine learning based software development effort estimation models. Information and Software Technology, Elsevier., 54, 41–59. https://doi.org/10.1016/j.infsof.2011.09.002.
Witten, I. H., & Frank, E. (2005). Data Mining. Second Edition, Elsevier: Practical Machine Learning Tools and Techniques.
Xiaochun, Z., Bo, Z., Fan, W., Yi, Q., Lu, C. (2008). Estimate Test Execution Effort at an Early Stage: An Empirical Study. In: IEEE International Conference on Cyberworlds, pp. 195–200, https://doi.org/10.1109/CW.2008.34.
Yáñez-Márquez, C., López-Yáñez, I., Aldape-Pérez, M., Camacho Nieto, O., Argüelles-Cruz, J. A., & Villuendas-Rey, Y. (2018). Theoretical foundations for the alpha-beta associative memories: 10 years of derived extensions, models, and applications. Neural Processing Letters, Elsevier, 48(2), 811–847. https://doi.org/10.1007/s11063-017-9768-2).
Yenigun, H., Yevtushenko, N., & Cavalli, A. R. (2019). Guest editorial: Special issue on testing software and systems. Software Quality Journal, Springer, 27, 497–499. https://doi.org/10.1007/s11219.
Acknowledgments
I would like to thank CUCEA of the Universidad de Guadalajara, Jalisco, México; Programa para el Desarrollo Profesional Docente (PRODEP), as well as to Consejo Nacional de Ciencia y Tecnología (Conacyt). In addition, I appreciate the help of the Ph.D. Student Felipe Belmont Polanco for his collaboration in the GP algorithm executions.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix
Appendix
Number of projects for effort analysis (TD, type of development; DP, development platform; LT, language type; CA, count approach; RL, resource level; NSP, number of software projects)
TD | DP | LT | CA | RL | NSP | TD | DP | LT | CA | RL | NSP |
---|---|---|---|---|---|---|---|---|---|---|---|
New | MF | 2GL | FISMA | 1 | 26 | Enhancement | MF | 2GL | IFPUG V4+ | 1 | 2 |
MR | 3GL | FISMA | 1 | 10 | 3GL | FISMA | 1 | 47 | |||
Multi | 3GL | FISMA | 1 | 37 | 3GL | IFPUG V4+ | 1 | 76 | |||
PC | 3GL | FISMA | 1 | 15 | 3GL | IFPUG V4+ | 2 | 3 | |||
MF | 4GL | FISMA | 1 | 4 | 3GL | FISMA | 4 | 3 | |||
MR | 4GL | FISMA | 1 | 4 | 3GL | IFPUG V4+ | 4 | 7 | |||
Multi | 4GL | FISMA | 1 | 10 | 4GL | FISMA | 1 | 6 | |||
PC | 4GL | FISMA | 1 | 2 | 4GL | IFPUG V4+ | 1 | 8 | |||
MR | ApG | FISMA | 1 | 4 | 4GL | IFPUG V4+ | 4 | 2 | |||
Multi | ApG | FISMA | 1 | 2 | ApG | FISMA | 1 | 4 | |||
MF | 3GL | FISMA | 4 | 1 | ApG | IFPUG V4+ | 1 | 53 | |||
PC | 3GL | FISMA | 4 | 2 | ApG | IFPUG V4+ | 2 | 1 | |||
MF | 2GL | IFPUG V4+ | 1 | 2 | MR | 3GL | FISMA | 1 | 4 | ||
MF | 3GL | IFPUG V4+ | 1 | 51 | 3GL | IFPUG V4+ | 1 | 29 | |||
MR | 3GL | IFPUG V4+ | 1 | 9 | 3GL | NESMA | 2 | 1 | |||
Multi | 3GL | IFPUG V4+ | 1 | 1 | 4GL | FISMA | 1 | 1 | |||
PC | 3GL | IFPUG V4+ | 1 | 21 | 4GL | IFPUG V4+ | 1 | 39 | |||
MF | 4GL | IFPUG V4+ | 1 | 12 | ApG | FISMA | 1 | 1 | |||
MR | 4GL | IFPUG V4+ | 1 | 8 | Multi | 3GL | FISMA | 1 | 13 | ||
Multi | 4GL | IFPUG V4+ | 1 | 3 | 3GL | IFPUG V4+ | 1 | 20 | |||
PC | 4GL | IFPUG V4+ | 1 | 26 | 3GL | NESMA | 4 | 1 | |||
MF | ApG | IFPUG V4+ | 1 | 2 | 4GL | FISMA | 1 | 4 | |||
MF | 3GL | IFPUG V4+ | 2 | 4 | 4GL | IFPUG V4+ | 4 | 1 | |||
MR | 3GL | IFPUG V4+ | 2 | 3 | ApG | FISMA | 1 | 9 | |||
MF | 4GL | IFPUG V4+ | 2 | 2 | PC | 3GL | IFPUG V4+ | 1 | 54 | ||
MR | 4GL | IFPUG V4+ | 2 | 1 | 3GL | IFPUG V4+ | 4 | 4 | |||
PC | 4GL | IFPUG V4+ | 2 | 2 | 4GL | IFPUG V4+ | 1 | 34 | |||
MF | ApG | IFPUG V4+ | 2 | 2 | 4GL | IFPUG V4+ | 4 | 1 | |||
MR | 3GL | IFPUG V4+ | 3 | 1 | |||||||
PC | 4GL | IFPUG V4+ | 3 | 1 | |||||||
MF | 3GL | IFPUG V4+ | 4 | 3 | |||||||
MR | 3GL | IFPUG V4+ | 4 | 1 | |||||||
Multi | 3GL | IFPUG V4+ | 4 | 2 | |||||||
PC | 3GL | IFPUG V4+ | 4 | 9 | |||||||
MF | 4GL | IFPUG V4+ | 4 | 1 | |||||||
MR | 4GL | IFPUG V4+ | 4 | 2 | |||||||
Multi | 4GL | IFPUG V4+ | 4 | 5 | |||||||
PC | 4GL | IFPUG V4+ | 4 | 9 | |||||||
MF | ApG | IFPUG V4+ | 4 | 1 |
Rights and permissions
About this article
Cite this article
López-Martín, C. Machine learning techniques for software testing effort prediction. Software Qual J 30, 65–100 (2022). https://doi.org/10.1007/s11219-020-09545-8
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s11219-020-09545-8