Skip to main content

Evaluation of Three Methods to Predict Project Success: A Case Study

  • Conference paper

Part of the book series: Lecture Notes in Computer Science ((LNPSE,volume 3547))

Abstract

To increase the likelihood for software project success, it is important to be able to identify the drivers of success. This paper compares three methods to identify similar projects with the objective to predict project success. The hypothesis is that projects with similar characteristics are likely to have the same outcome in terms of success. Two of the methods are based on identifying similar projects using all available information. The first method of these aims at identifying the most similar project. The second method identifies a group of projects as most similar. Finally, the third method pinpoints some key characteristics to identify project similarity. Our measure of success for these identifications is whether project success for these projects identified as similar is the same. The comparison between methods is done in a case study with 46 projects with varying characteristics. The paper evaluates the performance of each method with regards to its ability to predict project success. The method using key drivers of project success is superior to the others in the case study. Thus, it is concluded that it is important for software developing organizations to identify its key project characteristics to improve its control over project success.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. von Mayrhauser, A., Wohlin, C., Ohlsson, M.C.: Assessing and Understanding Efficiency and Success in Software Production. Empirical Software Engineering: An International Journal 5(2), 125–154 (2000)

    Article  MATH  Google Scholar 

  2. Kachigan, S.K.: Statistical Analysis – An Interdisciplinary Introduction to Univariate & Multivariate Methods. Radius Press (1986)

    Google Scholar 

  3. Shepperd, M.: Case-based Reasoning in Software Engineering. In: Aurum, A., Jeffrey, R., Wohlin, C., Handzic, M. (eds.) Managing Software Engineering Knowledge. Springer, Heidelberg (2003)

    Google Scholar 

  4. Basili, V.R., Caldiera, G., Rombach, H.D.: Experience Factory. In: Marciniak, J.J. (ed.) Encyclopedia of Software Engineering. John Wiley & Sons, Inc., Hoboken (2002)

    Google Scholar 

  5. Valett, J.D.: The (Mis)use of Subjective Process Measures in Software Engineering. In: Proc. Software Engineering Workshop, NASA/Goddard Space Flight Center, Greenbelt, Maryland, USA, pp. 161–165 (1993)

    Google Scholar 

  6. Gray, A.R., MacDonell, S.G., Shepperd, M.J.: Factors Systematically Associated with Errors in Subjective Estimates of Software Development Effort: The Stability of Expert Judgement. In: Proc. of the Sixth Int. Software Metrics Symposium, Boca Raton, Florida, USA, pp. 216–227 (1999)

    Google Scholar 

  7. Höst, M., Wohlin, C.: An Experimental Study of Individual Subjective Effort Estimations and Combinations of the Estimates. In: Proc. IEEE Int. Conf. on Software Engineering, Kyoto, Japan, pp. 332–339 (1998)

    Google Scholar 

  8. Ropponen, J., Lyytinen, K.: Components of Software Development Risk: How to Address Them? A Project Manager Survey. IEEE Trans. in Software Engineering 26(2), 98–112 (2000)

    Article  Google Scholar 

  9. Hughes, R.: Expert Judgement as an Estimation Method. Information and Software Technology 38, 67–75 (1996)

    Article  Google Scholar 

  10. Basili, V., Zelkowitz, M., McGarry, F., Page, J., Waligora, S., Pajerski, R.: SEL’s Software Process-Improvement Program. IEEE Software, 83–87 (November 1995)

    Google Scholar 

  11. Wohlin, C., Amschler Andrews, A.: Assessing Project Success using Subjective Evaluation Factors. Software Quality Journal 9(1), 43–70 (2000)

    Article  Google Scholar 

  12. Wohlin, C., von Mayrhauser, A., Höst, M., Regnell, B.: Subjective Evaluation as a Tool for Learning from Software Project Success. Information and Software Technology 42(14), 983–992 (2000)

    Article  Google Scholar 

  13. Myrtveit, I., Stensrud, E.: A Controlled Experiment to Assess the Benefits of Estimating with Analogy and Regression Models. IEEE Transactions on Software Engineering 25(4), 510–525 (1999)

    Article  Google Scholar 

  14. Shepperd, M., Kadoda, G.: Comparing Software Prediction Techniques Using Simulation. IEEE Transactions on Software Engineering 27(11), 1014–1022 (2001)

    Article  Google Scholar 

  15. Ohlsson, N., Zhao, M., Helander, M.: Application of Multivariate Analysis for Software Fault Prediction. Software Quality Journal 7(1), 51–66 (1998)

    Article  Google Scholar 

  16. Mair, C., Kadoda, G., Lefley, M., Phalp, K., Schofield, C., Shepperd, M., Webster, S.: An Investigation of Machine Learning Based Prediction Systems. Journal of Systems and Software 53(1), 23–29 (2000)

    Article  Google Scholar 

  17. Khoshgoftaar, T.M., Szabo, R.M.: Using Neural Networks to Predict Software Faults during Testing. IEEE Transaction on Reliability 45(3), 456–462 (1996)

    Article  Google Scholar 

  18. Shepperd, M.J., Schofield, C.: Estimating Software Project Effort Using Analogies. IEEE Transactions on Software Engineering 23(11), 736–743 (1997)

    Article  Google Scholar 

  19. Saaty, T.L., Vargas, L.G.: Models, Methods, Concepts & Applications of the Analytic Hierarchy Process. Kluwer Academic Publishers, Dordrecht (2001)

    Google Scholar 

  20. Shepperd, M., Cartwright, M.: Predicting with Sparse Data. IEEE Transactions on Software Engineering 27(11), 987–998 (2001)

    Article  Google Scholar 

  21. NASA-SEL: Software Engineering Laboratory Database Organization and Users Guide, Revision 2. Goddard Space Flight Center, Greenbelt, MD, USA, NASA-SEL Series, SEL-89-201 (1992)

    Google Scholar 

  22. El Emam, K., Wieczorek, I.: The Repeatability of Code Defect Classifications. In: Proc. the Ninth International Symposium on Software Reliability Engineering, pp. 322–333 (1998)

    Google Scholar 

  23. Altman, D.: Practical Statistics for Medical Research. Chapman and Hall, Boca Raton (1991)

    Google Scholar 

  24. Wohlin, C., Amschler Andrews, A.: A Case Study Approach to Evaluation of Three Methods to Predict Project Success. Technical Report, Blekinge Institute of Technology (2004), http://www.ipd.bth.se/cwo/TR-case.pdf

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Wohlin, C., Andrews, A.A. (2005). Evaluation of Three Methods to Predict Project Success: A Case Study. In: Bomarius, F., Komi-Sirviö, S. (eds) Product Focused Software Process Improvement. PROFES 2005. Lecture Notes in Computer Science, vol 3547. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11497455_31

Download citation

  • DOI: https://doi.org/10.1007/11497455_31

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-26200-8

  • Online ISBN: 978-3-540-31640-4

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics