Skip to main content
Log in

Developing artificial neural networks for safety critical systems

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

There are many performance based techniques that aim to improve the safety of neural networks for safety critical applications. However, many of these approaches provide inadequate forms of safety assurance required for certification. As a result, neural networks are typically restricted to advisory roles in safety-related applications. Neural networks have the ability to operate in unpredictable and changing environments. It is therefore desirable to certify them for highly-dependable roles in safety critical systems. This paper outlines the safety criteria which are safety requirements for the behaviour of neural networks. If enforced, the criteria can contribute to justifying the safety of ANN functional properties. Characteristics of potential neural network models are also outlined and are based upon representing knowledge in interpretable and understandable forms. The paper also presents a safety lifecycle for artificial neural networks. This lifecycle focuses on managing behaviour represented by neural networks and contributes to providing acceptable forms of safety assurance.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  1. Johnson E, Calise NAJ, Corban JE (2000) Adaptive guidance and control for autonomous launch vehicles. Georgia Institute of Technology & Guided Systems Technologies, Inc

  2. Sordo M, Buxton H, Watson D (2001) A hybrid approach to breast cancer diagnosis. School of Cognitive and Computing Sciences, University of Sussex

  3. Lisboa P (2001) Industrial use of safety-related artificial neural networks. Health Saf Exec pp 327

  4. IEC (1999) 61508: fundamental safety of electrical/electronic/programmable electronic safety related systems, International Electrotechnical Commission

  5. Partridge D (1997) Engineering multiversion reliability in neural networks—producing dependable systems. ERA Technology Ltd 97-0365

  6. Sharkey AJC, Sharkey NE, Gopinath OC (1995) Diversity, neural nets and safety critical applications. Computer Science, University of Sheffield

  7. Nabney I (2001) Validation of neural network medical systems. Workshop on regulatory issues in medical decision support, October 2001

  8. Saeks RE, Cox CJ, Sefic WJ, Graviss LP (1997) Verification and validation of a neural network flight control system. Accurate Automation Corporation, USA

    Google Scholar 

  9. Kurd Z (2002) Artificial neural networks in safety-critical applications. first year dissertation, Department of Computer Science, University of York

  10. Kurd Z, Kelly TP (2003) Establishing safety criteria for artificial neural networks. In: Seventh international conference on knowledge-based intelligent information & engineering systems (KES′03), Oxford, UK

  11. Kurd Z, Kelly TP (2002) Establishing safety criteria for artificial neural networks. Department of Computer Science, University of York, York, Internal Report 2002

  12. Kurd Z, Kelly TP (2003) Proposal for developing safety-critical artificial neural networks, Department of Computer Science, University of York, York, Internal Report, January 2003

  13. Kurd Z, Kelly TP (2003) Safety lifecycle for developing safety-critical artificial neural networks, 22nd International conference on computer safety, Reliability and security (SAFECOMP′03), 23–26, September 2003

  14. DARPA (1988) DARPA neural network study. AFCEA International Press

  15. Storey N (1996) Safety critical computer systems. Addison-Wesley, Harlow

    Google Scholar 

  16. Bell R, Reinert D (1993) Risk and system integrity concepts for safety-related control systems. Microprocess Microsy 17:13–15

    Google Scholar 

  17. SAE (1996) ARP 4761—Guidelines and methods for conducting the safety assessment process on civil airborne systems and equipment. The society for automotive engineers

  18. Bedford DF, Morgan G, Austin J (1996) Requirements for a standard certifying the use of artificial neural networks in safety critical applications. Proceedings of the international conference on artificial neural networks

  19. Kilimasaukas CC (1991) Neural nets tell why, Dr Dobbs′s, pp 16–24, April 1991

  20. Weaver RA, McDermid JA, Kelly TP (2002) Software safety arguments: towards a systematic categorisation of evidence. International System Safety Conference, Denver, CO

  21. Kelly TP (1998) Arguing safety—a systematic approach to managing safety cases. Ph.D. Thesis, Department of Computer Science, University of York

  22. Wen W, Callahan J, Napolitano M (1996) Towards developing verifiable neural network controller. Department of Aerospace Engineering, NASA/WVU Software Research Laboratory, West Virginia University

  23. Andrews R, Diederich J, Tickle A (1995) A survey and critique of techniques for extracting rules from trained artificial neural networks, Neurocomputing Research Centre, Queensland University of Technology

  24. G. Towell, Shavlik JW (1994) Knowledge-based artificial neural networks. Artif intell, pp 119–165

  25. Nauck D, Klawonn F, Kruse R (1993) Combining neural networks and fuzzy controllers, presented at FLAI′93, Linz, Austria

  26. Shavlik JW (1992) A framework for combining symbolic and neural learning. Computer Science Department, University of Wisconsin, Madison Technical Report 1123

  27. Kurd Z, Kelly TP (2003) Artificial neural networks in safety-critical applications. Thesis Proposal, Department of Computer Science, University of York

  28. Kurd Z, Kelly TP (2004) Defining a safety critical artificial neural network. Department of Computer Science, University of York, York, Internal Report

  29. Rodvold DM (1999) A software development process model for artificial neural networks in critical applications. Proceedings of the 1999 international conference on neural networks (IJCNN′99), Washington DC

  30. Nabney I, Paven MJS, Eldridge RC, Lee C (2000) Practical assessment of neural network applications, Aston University & Lloyd′s Register, UK

  31. Taha I, Ghosh J (1995) A hybrid intelligent architecture and its application to water reservoir control, Department of Electrical and Computer Engineering, University of Texas

  32. Rummelhart DE, Hinton GE, Williams RJ (1986) Learning representations by back-propagating errors. Nature 323:533–536

    Article  Google Scholar 

  33. Zwaag BJ, Spaanenburg L (2002) Analysis of neural network in terms of domain functions, presented at 3rd IEEE Benelux signal processing symposium (SPS-2002), Leuven, Belgium

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Zeshan Kurd.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Kurd, Z., Kelly, T. & Austin, J. Developing artificial neural networks for safety critical systems. Neural Comput & Applic 16, 11–19 (2007). https://doi.org/10.1007/s00521-006-0039-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-006-0039-9

Keywords