Abstract
This paper studies a general class of neural networks with time-varying delays and the neuron activations belong to the set of discontinuous monotone increasing functions. The discontinuities in the activations are an ideal model of the situation where the gain of the neuron amplifiers is very high. Because the delay in combination with high-gain nonlinearities is a particularly harmful source of potential instability, in the paper, conditions which ensure the global convergence of the neural network are derived.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Yi, Z., Heng, P.A., Vadakkepat, P.: Absolute Periodic and Absolute Stability of Delayed Neural Networks. IEEE Trans. Circuits and Systems I 49, 256–261 (2002)
Yi, Z., Heng, P.A., Leung, K.S.: Convergence Analysis of Celluar Neural Networks with Unbounded Delay. IEEE Trans. Circuits and Systems I 48, 680–687 (2001)
Yi, Z., Heng, P.A., Fu, A.W.C.: Estimate of Exponential Convergence Rate and Exponential Stability of Neural Networks with Unbounded Delay. IEEE Trans. Neural Networks 10, 1487–1493 (1999)
Liao, X., Wang, J.: Algebraic Criteria for Global Exponential Stability of Celluar Neural Networks with Multiple Time Delays. IEEE Trans. Circuits and Systems I 50, 268–285 (2003)
Qi, H., Qi, L.: Deriving Sufficient Conditions for Global Asymptotic Stability of Delayed Neural Networks via Nonsmooth Analysis. IEEE Trans. Neural Networks 15, 99–109 (2004)
Forti, M., Nistri, P.: Global Convergence of Neural Networks with Discontinuous Neural Activation. IEEE Trans. Circuits and Systems I 50, 1421–1435 (2003)
Forti, M., Nistri, P., Papini, D.: Global Exponential Stability and Global Convergence in Finite Time of Delayed Neural Networks with Infinite Gain. IEEE Trans. Neural Networks 16, 1449–1463 (2005)
Filippov, A.F.: Differential Equations with Discontinuous Right-Hand Side. Mathematics and its Applications (Soviet Series). Kluwer Academic Publishers, Boston (1988)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2007 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Zhang, L., Yi, Z. (2007). Global Exponential Convergence of Time-Varying Delayed Neural Networks with High Gain. In: Liu, D., Fei, S., Hou, ZG., Zhang, H., Sun, C. (eds) Advances in Neural Networks – ISNN 2007. ISNN 2007. Lecture Notes in Computer Science, vol 4491. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-72383-7_117
Download citation
DOI: https://doi.org/10.1007/978-3-540-72383-7_117
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-72382-0
Online ISBN: 978-3-540-72383-7
eBook Packages: Computer ScienceComputer Science (R0)