Skip to main content

Advertisement

Log in

Swarm intelligence based approach for efficient training of regressive neural networks

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

This work proposes an efficient approach to solve the problem of training a regressive neural network efficiently. Regressive networks are characterized by delay lines possibly in both the input and the output feedback. Each delay line is connected to the network with synaptic weights and thus increases the number of parameters that must be optimized by the training algorithm. Training algorithms such as the Levenberg–Marquardt, normally used to train neural networks, are prone to local minima entrapment, and for this reason, a strategy to initialize the training procedure correctly is needed. To solve this problem, the continuous flock of starling optimization algorithm, a highly explorative optimizer based on swarm intelligence, is used. The proposed approach is tested and validated on an experimental benchmark featuring a second-order nonlinear dynamic system.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

References

  1. Cybenko G (1989) Approximation by superpositions of a sigmoidal function. Math Control Signals Syst (MCSS) 2(4):303–314

    MathSciNet  MATH  Google Scholar 

  2. Hornik K, Stinchcombe M, White H (1989) Multilayer feedforward networks are universal approximators. Neural Netw 2(5):359–366

    MATH  Google Scholar 

  3. Whalen P, Brio M, Moloney JV (2015) Exponential time-differencing with embedded Runge–Kutta adaptive step control. J Comput Phys 280:579–601

    MathSciNet  MATH  Google Scholar 

  4. Butcher JC (2016) Numerical methods for ordinary differential equations. Wiley, New York

    MATH  Google Scholar 

  5. Khan K, Sahai A (2012) A comparison of BA, GA, PSO, BP and LM for training feed forward neural networks in e-learning context. Int J Intell Syst Appl 4(7):23

    Google Scholar 

  6. De Jesús Rubio J, Angelov P, Pacheco J (2011) Uniformly stable backpropagation algorithm to train a feedforward neural network. IEEE Trans Neural Netw 22(3):356–366

    Google Scholar 

  7. Williams RJ, Zipser D (1989) Experimental analysis of the real-time recurrent learning algorithm. Connect Sci 1(1):87–111

    Google Scholar 

  8. Steil JJ (2004) Backpropagation-decorrelation: online recurrent learning with O (N) complexity. In: 2004 IEEE international joint conference on neural networks. Proceedings, vol 2. IEEE, pp 843–848

  9. Williams RJ, Zipser D (1989) A learning algorithm for continually running fully recurrent neural networks. Neural Comput 1(2):270–280

    Google Scholar 

  10. Laudani A, Lozito GM, Riganti Fulginei F, Salvini A (2015) On training efficiency and computational costs of a feed forward neural network: a review. Comput Intell Neurosci 2015:83

    Google Scholar 

  11. Youssef A, Mohammed K, Nasar A (2012) A reconfigurable, generic and programmable feed forward neural network implementation in FPGA. In: 2012 UKSim 14th international conference on computer modelling and simulation (UKSim). IEEE, pp 9–13

  12. Hariprasath S, Prabakar TN (2012) FPGA implementation of multilayer feed forward neural network architecture using VHDL. In: 2012 international conference on computing, communication and applications (ICCCA). IEEE, pp 1–6

  13. Laudani A, Lozito GM, Fulginei FR, Salvini A (2014) An efficient architecture for floating point based MISO neural neworks on FPGA. In: 2014 UKSim-AMSS 16th international conference on computer modelling and simulation (UKSim). IEEE, pp 12–17

  14. Cardelli E, Faba A, Laudani A, Lozito GM, Fulginei FR, Salvini A (2016) Two-dimensional magnetic modeling of ferromagnetic materials by using a neural networks based hybrid approach. Physica B 486:106–110

    Google Scholar 

  15. Cardelli E, Faba A, Laudani A, Lozito GM, Fulginei FR, Salvini A (2016) A neural-FEM tool for the 2-D magnetic hysteresis modeling. Physica B 486:111–115

    Google Scholar 

  16. Ilonen J, Kamarainen JK, Lampinen J (2003) Differential evolution training algorithm for feed-forward neural networks. Neural Process Lett 17(1):93–105

    Google Scholar 

  17. Leung FHF, Lam HK, Ling SH, Tam PKS (2003) Tuning of the structure and parameters of a neural network using an improved genetic algorithm. IEEE Trans Neural Netw 14(1):79–88

    Google Scholar 

  18. Karaboga D, Akay B, Ozturk C (2007) Artificial bee colony (ABC) optimization algorithm for training feed-forward neural networks. MDAI 7:318–319

    Google Scholar 

  19. Zhang JR, Zhang J, Lok TM, Lyu MR (2007) A hybrid particle swarm optimization–back-propagation algorithm for feedforward neural network training. Appl Math Comput 185(2):1026–1037

    MATH  Google Scholar 

  20. Valian E, Mohanna S, Tavakoli S (2011) Improved cuckoo search algorithm for feedforward neural network training. Int J Artif Intell Appl 2(3):36–43

    Google Scholar 

  21. Laudani A, Fulginei FR, Salvini A (2013) Closed forms for the fully-connected continuous flock of starlings optimization algorithm. In: 2013 UKSim 15th international conference on computer modelling and simulation (UKSim). IEEE, pp 45–50

  22. Laudani A, Riganti Fulginei F, Salvini A, Schmid M, Conforto S (2013) CFSO3: a new supervised swarm-based optimization algorithm. Math Probl Eng. https://doi.org/10.1155/2013/560614

    Article  MATH  Google Scholar 

  23. Laudani A, Fulginei FR, Lozito GM, Salvini A (2014) Swarm/flock optimization algorithms as continuous dynamic systems. Appl Math Comput 243:670–683

    MathSciNet  MATH  Google Scholar 

  24. Schoukens JPN (2017) Three benchmarks addressing open challenges in nonlinear system identification. IFAC-PapersOnLine, 50(1):446–451. ISSN 2405-8963. https://doi.org/10.1016/j.ifacol.2017.08.071

  25. http://www.nonlinearbenchmark.org/FILES/BENCHMARKS/CASCADEDTANKS/CascadedTanksFiles.zip

  26. Lozito GM, Riganti Fulginei F, Salvini A (2015) On the generalization capabilities of the ten-parameter Jiles–Atherton model. Math Probl Eng. https://doi.org/10.1155/2015/715018

    Article  MathSciNet  MATH  Google Scholar 

  27. Coco S, Laudani A, Lozito GM, Pollicino G (2018) Effective permeability estimation of a composite magnetic shielding mortar by using swarm intelligence. Int J Appl Electromagn Mech 1–12 (Preprint)

  28. Laudani A, Fulginei FR, Salvini A (2015) TMS array coils optimization by means of CFSO. IEEE Trans Magn 51(3):1–4

    Google Scholar 

  29. Laudani A, Lozito GM, Fulginei FR, Salvini A (2016) FEM model identification for a vector hysteresis workbench. In: 2016 IEEE 2nd international forum on research and technologies for society and industry leveraging a better tomorrow (RTSI). IEEE, pp 1–5

  30. Coco S, Laudani A, Fulginei FR, Salvini A (2012) Accurate design of Helmholtz coils for ELF Bioelectromagnetic interaction by means of continuous FSO. Int J Appl Electromagn Mech 39(1–4):651–656

    Google Scholar 

  31. Piotrowski AP, Napiorkowski JJ (2011) Optimizing neural networks for river flow forecasting–evolutionary computation methods versus the Levenberg–Marquardt approach. J Hydrol 407(1–4):12–27

    Google Scholar 

  32. Wolpert DH, Macready WG (1997) No free lunch theorems for optimization. IEEE Trans Evol Comput 1(1):67–82

    Google Scholar 

  33. Ceperic V, Baric A (2014) Reducing complexity of echo state networks with sparse linear regression algorithms. In: 2014 UKSim-AMSS 16th international conference on computer modelling and simulation (UKSim), pp 26–31

  34. Jaeger H (2003) Adaptive nonlinear system identification with echo state networks. In: Advances in neural information processing systems, pp 609–616. ISBN: 0262025507;978-026202550-8

  35. Atiya Amir F, Parlos Alexander G (2000) New results on recurrent network training: unifying the algorithms and accelerating convergence. IEEE Trans Neural Netw 11(3):697–709

    Google Scholar 

  36. Verstraeten D, Schrauwen B, d’Haene M, Stroobandt D (2007) An experimental unification of reservoir computing methods. Neural Netw 20(3):391–403

    MATH  Google Scholar 

  37. Rodan A, Tino P (2011) Minimum complexity echo state network. IEEE Trans Neural Netw 22(1):131–144

    Google Scholar 

  38. Gensler A, Henze J, Sick B, Raabe N (2016) Deep learning for solar power forecasting—an approach using AutoEncoder and LSTM neural networks. In: 2016 IEEE international conference on systems, man, and cybernetics (SMC). IEEE, pp 002858–002865

  39. http://www.sea.uniroma3.it/elettrotecnica/tools.php

  40. Łukasik S, Kowalski PA (2014) Fully informed swarm optimization algorithms: basic concepts, variants and experimental evaluation. In: 2014 Federated conference on computer science and information systems (FedCSIS). IEEE, pp 155–61

  41. Wang S, Phillips P, Yang J, Sun P, Zhang Y (2016) Magnetic resonance brain classification by a novel binary particle swarm optimization with mutation and time-varying acceleration coefficients. Biomed Eng (Biomedizinische Technik) 61(4):431–441

    Google Scholar 

  42. Zhang Y, Wang S, Phillips P, Dong Z, Ji G, Yang J (2015) Detection of Alzheimer’s disease and mild cognitive impairment based on structural volumetric MR images using 3D-DWT and WTA-KSVM trained by PSOTVAC. Biomed Signal Process Control 21:58–73

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Gabriele Maria Lozito.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix: Comparison against classic GA and PSO

Appendix: Comparison against classic GA and PSO

The present appendix reports a brief comparison against classic global optimization techniques, in place of the CFSO, for the initialization of the search space to be investigated by the LM. In particular, the genetic algorithm (GA) and the classic PSO were used. Both the PSO and the GA were used considering a set of 10 agents (i.e., 10 particles for the PSO, 10 individuals for the GA). For the GA, an intermediate crossover rule was used with a 10% proportional mutation probability. No elitism was implemented. The PSO was directly derived from the FSO by forcing to 0 the H terms. Analogously to the tests described in Sect. 6.2, 300 iterations were considered: 200 for the global optimization algorithm (PSO or GA) and 100 for the refinement via LM. The test was done only on the optimal sizing parameters of the NARX; thus, N = 7, DU = 3 and DX = 1. For both algorithms GA + LM and PSO + LM, the test was repeated 150 times. Average RMSE for the GA + LM is 0.05981, and average RMSE for the PSO + LM is 0.07005. Both the methods outperform random initialization of the LM, with a distinct advantage of the GA over the PSO. Still, the PSO used is the simplest version ever formulated, and several variations [40,41,42] exist that can give much better optimization capabilities. Results are summarized in Table 5, and a time response of the system is shown in Fig. 6. As can be seen from the figure, GA and CFSO have very similar performance, and PSO shows a small accuracy degradation.

Table 5 Comparative analysis against classic meta-heuristics
Fig. 6
figure 6

Time response for the double-tank dynamic system (bottom, black), the optimally trained neural network by CFSO + LM (bottom, red), the optimally trained neural network by GA + LM (bottom, blue), the optimally trained neural network by PSO + LM (bottom, magenta), given a time-variable input (top, pale blue). Simulations are performed on the validation set (color figure online)

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Lozito, G.M., Salvini, A. Swarm intelligence based approach for efficient training of regressive neural networks. Neural Comput & Applic 32, 10693–10704 (2020). https://doi.org/10.1007/s00521-019-04606-x

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-019-04606-x

Keywords