Skip to main content
Log in

A framework for intelligent test data generation

  • Published:
Journal of Intelligent and Robotic Systems Aims and scope Submit manuscript

Abstract

Test data generation using traditional software testing methods generally requires considerable manual effort and generates only a limited number of test cases before the amount of time expanded becomes unacceptably large. A rule-based framework that will automatically generate test data to achieve maximal branch coverage is presented. The design and discovery of rules used to generate meaningful test cases are also described. The rule-based approach allows this framework to be extended to include additional testing requirements and test case generation knowledge.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Similar content being viewed by others

References

  1. Adrion, W.R. et al., ‘Validation, verification, and testing of computer software’, ACM Computing Surveys, 14 (1982).

  2. Brown, D.B., The Development of a Program Analysis Environment for Ada-Phase II (Task I)-NASA Six-month-report, Department of Computer Science and Engineering, Auburn University, December (1989).

  3. Brown, D.B., The Development of a Program Analysis Environment for Ada-Phase II (Task I)-NASA Annual Report, Department of Computer Science and Engineering, Auburn University, August (1990).

  4. CLIPS Reference Manual, Version 4.1, Artificial Intelligence Section, Johnson Space Center, NASA, September 1987.

  5. Cross, J.H., II, Morrison, K.I., May, C.H., and Waddel, K.C., A Graphically Oriented Specification Language for Automatic Code Generation (Phase 1), NASA Annual Report, Department of Computer Science and Engineering, Auburn University, August (1989).

  6. Deason, W.H., Brown, D.B., Chang, K.-H. and Cross, J.H., II, ‘A rule-based software test data generator’, IEEE Trans. Knowledge Data Engrg., March 1991.

  7. DeMillo, R.A., Lipton, R.J., and Sayward, F.G., ‘Hints on test data selection: help for the practising programmer’, IEEE Computer, 11, (1978).

  8. DIANA Interface Package Manual, Verdix Corp., CA. (1990).

  9. Duran, J.W. and Ntafos, S., ‘A report on random testing’, Proc. 5th Int. Conf. on Software Engineering, March (1981).

  10. Howden, W.E., ‘A functional approach to program testing and analysis’, IEEE Trans. Software Engrg. SE-12, (1986).

  11. Prather, R.E. and Myers, P. Jr., ‘The path prefix software testing strategy’, IEEE Trans. Software Engrg. SE-13 (1987).

  12. Ramamoorthy, C.V. and Ho, S.F., ‘Testing large software with automated software evaluation systems’, IEEE Trans. Software Engrg. SE-1 (1975).

  13. Vouk, M.A., McAllister, D.F. and Tai, T.C., ‘An experimental evaluation of the effectiveness of random testing of fault-tolerant software’, Proc. IEEE Workshop on Software Testing (1986).

Download references

Author information

Authors and Affiliations

Authors

Additional information

This work was supported in part by George C. Marshall Space Flight Center, NASA/MSFC, AL 35812 (NASA-NCC8-14).

Rights and permissions

Reprints and permissions

About this article

Cite this article

Chang, KH., Cross, J.H., Carlisle, W.H. et al. A framework for intelligent test data generation. J Intell Robot Syst 5, 147–165 (1992). https://doi.org/10.1007/BF00444293

Download citation

  • Received:

  • Accepted:

  • Issue Date:

  • DOI: https://doi.org/10.1007/BF00444293

Key words

Navigation