Abstract
There are many performance based techniques that aim to improve the safety of neural networks for safety critical applications. However, many of these approaches provide inadequate forms of safety assurance required for certification. As a result, neural networks are typically restricted to advisory roles in safety-related applications. Neural networks have the ability to operate in unpredictable and changing environments. It is therefore desirable to certify them for highly-dependable roles in safety critical systems. This paper outlines the safety criteria which are safety requirements for the behaviour of neural networks. If enforced, the criteria can contribute to justifying the safety of ANN functional properties. Characteristics of potential neural network models are also outlined and are based upon representing knowledge in interpretable and understandable forms. The paper also presents a safety lifecycle for artificial neural networks. This lifecycle focuses on managing behaviour represented by neural networks and contributes to providing acceptable forms of safety assurance.



Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Johnson E, Calise NAJ, Corban JE (2000) Adaptive guidance and control for autonomous launch vehicles. Georgia Institute of Technology & Guided Systems Technologies, Inc
Sordo M, Buxton H, Watson D (2001) A hybrid approach to breast cancer diagnosis. School of Cognitive and Computing Sciences, University of Sussex
Lisboa P (2001) Industrial use of safety-related artificial neural networks. Health Saf Exec pp 327
IEC (1999) 61508: fundamental safety of electrical/electronic/programmable electronic safety related systems, International Electrotechnical Commission
Partridge D (1997) Engineering multiversion reliability in neural networks—producing dependable systems. ERA Technology Ltd 97-0365
Sharkey AJC, Sharkey NE, Gopinath OC (1995) Diversity, neural nets and safety critical applications. Computer Science, University of Sheffield
Nabney I (2001) Validation of neural network medical systems. Workshop on regulatory issues in medical decision support, October 2001
Saeks RE, Cox CJ, Sefic WJ, Graviss LP (1997) Verification and validation of a neural network flight control system. Accurate Automation Corporation, USA
Kurd Z (2002) Artificial neural networks in safety-critical applications. first year dissertation, Department of Computer Science, University of York
Kurd Z, Kelly TP (2003) Establishing safety criteria for artificial neural networks. In: Seventh international conference on knowledge-based intelligent information & engineering systems (KES′03), Oxford, UK
Kurd Z, Kelly TP (2002) Establishing safety criteria for artificial neural networks. Department of Computer Science, University of York, York, Internal Report 2002
Kurd Z, Kelly TP (2003) Proposal for developing safety-critical artificial neural networks, Department of Computer Science, University of York, York, Internal Report, January 2003
Kurd Z, Kelly TP (2003) Safety lifecycle for developing safety-critical artificial neural networks, 22nd International conference on computer safety, Reliability and security (SAFECOMP′03), 23–26, September 2003
DARPA (1988) DARPA neural network study. AFCEA International Press
Storey N (1996) Safety critical computer systems. Addison-Wesley, Harlow
Bell R, Reinert D (1993) Risk and system integrity concepts for safety-related control systems. Microprocess Microsy 17:13–15
SAE (1996) ARP 4761—Guidelines and methods for conducting the safety assessment process on civil airborne systems and equipment. The society for automotive engineers
Bedford DF, Morgan G, Austin J (1996) Requirements for a standard certifying the use of artificial neural networks in safety critical applications. Proceedings of the international conference on artificial neural networks
Kilimasaukas CC (1991) Neural nets tell why, Dr Dobbs′s, pp 16–24, April 1991
Weaver RA, McDermid JA, Kelly TP (2002) Software safety arguments: towards a systematic categorisation of evidence. International System Safety Conference, Denver, CO
Kelly TP (1998) Arguing safety—a systematic approach to managing safety cases. Ph.D. Thesis, Department of Computer Science, University of York
Wen W, Callahan J, Napolitano M (1996) Towards developing verifiable neural network controller. Department of Aerospace Engineering, NASA/WVU Software Research Laboratory, West Virginia University
Andrews R, Diederich J, Tickle A (1995) A survey and critique of techniques for extracting rules from trained artificial neural networks, Neurocomputing Research Centre, Queensland University of Technology
G. Towell, Shavlik JW (1994) Knowledge-based artificial neural networks. Artif intell, pp 119–165
Nauck D, Klawonn F, Kruse R (1993) Combining neural networks and fuzzy controllers, presented at FLAI′93, Linz, Austria
Shavlik JW (1992) A framework for combining symbolic and neural learning. Computer Science Department, University of Wisconsin, Madison Technical Report 1123
Kurd Z, Kelly TP (2003) Artificial neural networks in safety-critical applications. Thesis Proposal, Department of Computer Science, University of York
Kurd Z, Kelly TP (2004) Defining a safety critical artificial neural network. Department of Computer Science, University of York, York, Internal Report
Rodvold DM (1999) A software development process model for artificial neural networks in critical applications. Proceedings of the 1999 international conference on neural networks (IJCNN′99), Washington DC
Nabney I, Paven MJS, Eldridge RC, Lee C (2000) Practical assessment of neural network applications, Aston University & Lloyd′s Register, UK
Taha I, Ghosh J (1995) A hybrid intelligent architecture and its application to water reservoir control, Department of Electrical and Computer Engineering, University of Texas
Rummelhart DE, Hinton GE, Williams RJ (1986) Learning representations by back-propagating errors. Nature 323:533–536
Zwaag BJ, Spaanenburg L (2002) Analysis of neural network in terms of domain functions, presented at 3rd IEEE Benelux signal processing symposium (SPS-2002), Leuven, Belgium
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Kurd, Z., Kelly, T. & Austin, J. Developing artificial neural networks for safety critical systems. Neural Comput & Applic 16, 11–19 (2007). https://doi.org/10.1007/s00521-006-0039-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-006-0039-9