Skip to main content
Log in

Finite-Time Stability and Its Application for Solving Time-Varying Sylvester Equation by Recurrent Neural Network

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

This paper investigates finite-time stability and its application for solving time-varying Sylvester equation by recurrent neural network. Firstly, a new finite-time stability criterion is given and a less conservative upper bound of the convergence time is also derived. Secondly, a sign-bi-power activation function with a linear term is presented for the recurrent neural network. The estimation of the upper bound of the convergence time is more less conservative. Thirdly, it is proposed a tunable activation function with three tunable positive parameters for the recurrent neural network. These parameters are not only helpful to reduce conservatism of the upper bound of the convergence time, accelerate convergence but also reduce sensitivity to additive noise. The effectiveness of our methods is shown by both theoretical analysis and numerical simulations.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7
Fig. 8
Fig. 9
Fig. 10
Fig. 11
Fig. 12
Fig. 13
Fig. 14
Fig. 15
Fig. 16
Fig. 17

Similar content being viewed by others

References

  1. Skowronski MD, Harris JG (2007) Noise-robust automatic speech recognition using a predictive echo state network. IEEE Trans Audio Speech Lang Process 15(5):1724–1730

    Article  Google Scholar 

  2. Burrows N, Niranjan M (1994) The use of recurrent neural networks for classification. In: IEEE workshop on neural networks for signal processing IV, Glasgow, pp 117–125

  3. Husken M, Stagge P (2003) Recurrent neural networks for time series classification. Neurocomputing 50:223–235

    Article  Google Scholar 

  4. Li S, Chen S, Liu B, Li Y, Liang Y (2012) Decentralized kinematic control of a class of collaborayive redundant manipulators via recurrent neural networks. Neurocomputing 91:1–10

    Article  Google Scholar 

  5. Lin S, Li Y, Liu B (2012) Model-free control of Lorenz chaos using an approximate optimal control strategy. Commun Nonlinear Sci Numer Simul 12(7):4891–4900

    MathSciNet  Google Scholar 

  6. Smith KA (1999) Neural networks for combinatorial optimization: a review of more than a decade of research. INFORMS J Comput 11:15–34

    Article  MathSciNet  MATH  Google Scholar 

  7. Li S, Lou Y, Liu B (2012) Bluetooth aided mobile phone localization: a nonlinear neural circuit approach. ACM Trans Embed Comput Syst

  8. Hopfield JJ (1984) Neurons with graded response have collective computational properties like those of two-state neurons. Proc Natl Acad Sci USA 81(10):3088–3092

    Article  Google Scholar 

  9. Ding F, Chen T (2005) Gradient based iterative algorithms for solving a class of matrix equations. IEEE Trans Auto Control 50(8):1216–1221

    Article  MathSciNet  Google Scholar 

  10. Zhou B, Li ZY, Duan GR, Wang Y (2009) Least squares solutions to general coupled Sylvester matrix equations. J Comput Appl Math 224:759–776

    Article  MathSciNet  MATH  Google Scholar 

  11. Manherz R, Jordan B, Hakimi S (1968) Analog methods for computation of the generalized inverse. IEEE Trans Auto Control 13(5):582–585

    Article  Google Scholar 

  12. Jang JS, Lee SY, Shin SY (1987) An optimization network for matrix inversion. In: Proceedings of NIPS87, Graz, pp 397–401

  13. Zhang Y, Ge SS (2005) Design and analysis of a general recurrent neural network model for time-varying matrix inversion. IEEE Trans Neural Netw 16(6):1477–1490

    Article  Google Scholar 

  14. Xiao L, Zhang Y (2011) Zhang neural network versus gradient neural network for solving time-varying linear inequalities. IEEE Trans Neural Netw 22(10):1676–1684

    Article  MathSciNet  Google Scholar 

  15. Zhang Y, Ma W, Cai B (2009) From Zhang neural network to Newton iteration for matrix inversion. IEEE Trans Circuits Syst I 56(7):1405–1415

    Article  MathSciNet  Google Scholar 

  16. Zhang Y, Yi C (2011) Zhang neural networks & neural-dynamic method. Nova Science Publishers Inc., New York

    Google Scholar 

  17. Zhang Y, Yi C, Guo D, Zheng J (2011) Comparison on Zhang neural dynamics and gradient-based neural dynamics for online solution of nonlinear time-varying equation. Neural Comput Appl 20:1–7

    Article  Google Scholar 

  18. Bartels RH, Stewart GW (1972) Solution of the matrix equation \(AX + XB = C\). Commun ACM 15:820–826

    Article  Google Scholar 

  19. Li S, Chen SF, Li B (2013) Accelerating a recurrent neural network to finite-time convergence for solving time-varying Sylvester equation by using a sign-bi-power activation function. Neural Process Lett 37:189–205

    Article  Google Scholar 

  20. Bhat SP, Bernstein DS (2000) Finite-time stability of continuous autonomous systems. SIAM J Control Optim 38:751–766

    Article  MathSciNet  MATH  Google Scholar 

  21. Shen YJ, Huang YH (2012) Global finite-time stabilisation for a class of nonlinear systems. Int J Syst Sci 43(1):73–78

    Article  MathSciNet  MATH  Google Scholar 

  22. Wang WP, Peng HP, Li LX, Xiao JH, Yang YX (2013) Finite-time function projective synchronization in complex multi-links networks with time-varying delay. Neural Process Lett. doi:10.1007/s11063-013-9335-4

  23. Bertsekas DP, Tsitsiklis JN (1996) Neuro-dynamic programming, optimization and neural computation series, 3. Athena Scientific, Belmont

    Google Scholar 

  24. Kamruzzaman J, Aziz SM (2002) A note on activation function in multilayer feed forward learning. In: Proceedings of the 2002 international joint conference on neural networks 2002, IJCNN 02, vol 1, Honolulu, pp 519–523

  25. Gao WB, Huang JC (1993) Variable structure control of nonlinear systems: a new approach. IEEE Trans Ind Electron 40:45–55

    Article  Google Scholar 

  26. Wang WP, Li LX, Peng HB, Xiao JH, Yang YX (2014) Synchronization control of memristor-based recurrent neural networks with perturbations. Neural Netw 53:8–14

    Article  MATH  Google Scholar 

  27. Li LX, Kurths Jrgen (2013) Exponentially asymptotic synchronization of uncertain complex time-delay dynamical networks. Eur Phys J B 86:125. doi:10.1140/epjb/e2013-30517-6

    Article  MathSciNet  Google Scholar 

  28. Myers RH (1994) Classical and modern regression with applications, 2nd edn. PWS-KENT Publishing Company, Boston

    Google Scholar 

  29. Buss SR (2004) Introduction to inverse kinematics with Jacobian transpose. Pseudo inverse and damped least squares methods. Technical Report. University of California, San Diego

    Google Scholar 

  30. Stephen LC, Carl DM (2009) Generalized inverses of linear transformations. Society for Industrial and Applied Mathematics, Philadephia

    MATH  Google Scholar 

  31. Banerjee KS (1980) Generalized inverse of matrices and its applications. Technometrics 15(1):471–512

    Google Scholar 

Download references

Acknowledgments

This work was supported by the National Science Foundation of China (No. 61374028, 61273183, 51177088), the Grant National Science Foundation of Hubei Provincial (2013CFA05 0), the Graduate Scientific Research Foundation of China Three Gorges University (2014PY069).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yanjun Shen.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Shen, Y., Miao, P., Huang, Y. et al. Finite-Time Stability and Its Application for Solving Time-Varying Sylvester Equation by Recurrent Neural Network. Neural Process Lett 42, 763–784 (2015). https://doi.org/10.1007/s11063-014-9397-y

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11063-014-9397-y

Keywords

Navigation