Abstract
This work proposes an efficient approach to solve the problem of training a regressive neural network efficiently. Regressive networks are characterized by delay lines possibly in both the input and the output feedback. Each delay line is connected to the network with synaptic weights and thus increases the number of parameters that must be optimized by the training algorithm. Training algorithms such as the Levenberg–Marquardt, normally used to train neural networks, are prone to local minima entrapment, and for this reason, a strategy to initialize the training procedure correctly is needed. To solve this problem, the continuous flock of starling optimization algorithm, a highly explorative optimizer based on swarm intelligence, is used. The proposed approach is tested and validated on an experimental benchmark featuring a second-order nonlinear dynamic system.





Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Cybenko G (1989) Approximation by superpositions of a sigmoidal function. Math Control Signals Syst (MCSS) 2(4):303–314
Hornik K, Stinchcombe M, White H (1989) Multilayer feedforward networks are universal approximators. Neural Netw 2(5):359–366
Whalen P, Brio M, Moloney JV (2015) Exponential time-differencing with embedded Runge–Kutta adaptive step control. J Comput Phys 280:579–601
Butcher JC (2016) Numerical methods for ordinary differential equations. Wiley, New York
Khan K, Sahai A (2012) A comparison of BA, GA, PSO, BP and LM for training feed forward neural networks in e-learning context. Int J Intell Syst Appl 4(7):23
De Jesús Rubio J, Angelov P, Pacheco J (2011) Uniformly stable backpropagation algorithm to train a feedforward neural network. IEEE Trans Neural Netw 22(3):356–366
Williams RJ, Zipser D (1989) Experimental analysis of the real-time recurrent learning algorithm. Connect Sci 1(1):87–111
Steil JJ (2004) Backpropagation-decorrelation: online recurrent learning with O (N) complexity. In: 2004 IEEE international joint conference on neural networks. Proceedings, vol 2. IEEE, pp 843–848
Williams RJ, Zipser D (1989) A learning algorithm for continually running fully recurrent neural networks. Neural Comput 1(2):270–280
Laudani A, Lozito GM, Riganti Fulginei F, Salvini A (2015) On training efficiency and computational costs of a feed forward neural network: a review. Comput Intell Neurosci 2015:83
Youssef A, Mohammed K, Nasar A (2012) A reconfigurable, generic and programmable feed forward neural network implementation in FPGA. In: 2012 UKSim 14th international conference on computer modelling and simulation (UKSim). IEEE, pp 9–13
Hariprasath S, Prabakar TN (2012) FPGA implementation of multilayer feed forward neural network architecture using VHDL. In: 2012 international conference on computing, communication and applications (ICCCA). IEEE, pp 1–6
Laudani A, Lozito GM, Fulginei FR, Salvini A (2014) An efficient architecture for floating point based MISO neural neworks on FPGA. In: 2014 UKSim-AMSS 16th international conference on computer modelling and simulation (UKSim). IEEE, pp 12–17
Cardelli E, Faba A, Laudani A, Lozito GM, Fulginei FR, Salvini A (2016) Two-dimensional magnetic modeling of ferromagnetic materials by using a neural networks based hybrid approach. Physica B 486:106–110
Cardelli E, Faba A, Laudani A, Lozito GM, Fulginei FR, Salvini A (2016) A neural-FEM tool for the 2-D magnetic hysteresis modeling. Physica B 486:111–115
Ilonen J, Kamarainen JK, Lampinen J (2003) Differential evolution training algorithm for feed-forward neural networks. Neural Process Lett 17(1):93–105
Leung FHF, Lam HK, Ling SH, Tam PKS (2003) Tuning of the structure and parameters of a neural network using an improved genetic algorithm. IEEE Trans Neural Netw 14(1):79–88
Karaboga D, Akay B, Ozturk C (2007) Artificial bee colony (ABC) optimization algorithm for training feed-forward neural networks. MDAI 7:318–319
Zhang JR, Zhang J, Lok TM, Lyu MR (2007) A hybrid particle swarm optimization–back-propagation algorithm for feedforward neural network training. Appl Math Comput 185(2):1026–1037
Valian E, Mohanna S, Tavakoli S (2011) Improved cuckoo search algorithm for feedforward neural network training. Int J Artif Intell Appl 2(3):36–43
Laudani A, Fulginei FR, Salvini A (2013) Closed forms for the fully-connected continuous flock of starlings optimization algorithm. In: 2013 UKSim 15th international conference on computer modelling and simulation (UKSim). IEEE, pp 45–50
Laudani A, Riganti Fulginei F, Salvini A, Schmid M, Conforto S (2013) CFSO3: a new supervised swarm-based optimization algorithm. Math Probl Eng. https://doi.org/10.1155/2013/560614
Laudani A, Fulginei FR, Lozito GM, Salvini A (2014) Swarm/flock optimization algorithms as continuous dynamic systems. Appl Math Comput 243:670–683
Schoukens JPN (2017) Three benchmarks addressing open challenges in nonlinear system identification. IFAC-PapersOnLine, 50(1):446–451. ISSN 2405-8963. https://doi.org/10.1016/j.ifacol.2017.08.071
http://www.nonlinearbenchmark.org/FILES/BENCHMARKS/CASCADEDTANKS/CascadedTanksFiles.zip
Lozito GM, Riganti Fulginei F, Salvini A (2015) On the generalization capabilities of the ten-parameter Jiles–Atherton model. Math Probl Eng. https://doi.org/10.1155/2015/715018
Coco S, Laudani A, Lozito GM, Pollicino G (2018) Effective permeability estimation of a composite magnetic shielding mortar by using swarm intelligence. Int J Appl Electromagn Mech 1–12 (Preprint)
Laudani A, Fulginei FR, Salvini A (2015) TMS array coils optimization by means of CFSO. IEEE Trans Magn 51(3):1–4
Laudani A, Lozito GM, Fulginei FR, Salvini A (2016) FEM model identification for a vector hysteresis workbench. In: 2016 IEEE 2nd international forum on research and technologies for society and industry leveraging a better tomorrow (RTSI). IEEE, pp 1–5
Coco S, Laudani A, Fulginei FR, Salvini A (2012) Accurate design of Helmholtz coils for ELF Bioelectromagnetic interaction by means of continuous FSO. Int J Appl Electromagn Mech 39(1–4):651–656
Piotrowski AP, Napiorkowski JJ (2011) Optimizing neural networks for river flow forecasting–evolutionary computation methods versus the Levenberg–Marquardt approach. J Hydrol 407(1–4):12–27
Wolpert DH, Macready WG (1997) No free lunch theorems for optimization. IEEE Trans Evol Comput 1(1):67–82
Ceperic V, Baric A (2014) Reducing complexity of echo state networks with sparse linear regression algorithms. In: 2014 UKSim-AMSS 16th international conference on computer modelling and simulation (UKSim), pp 26–31
Jaeger H (2003) Adaptive nonlinear system identification with echo state networks. In: Advances in neural information processing systems, pp 609–616. ISBN: 0262025507;978-026202550-8
Atiya Amir F, Parlos Alexander G (2000) New results on recurrent network training: unifying the algorithms and accelerating convergence. IEEE Trans Neural Netw 11(3):697–709
Verstraeten D, Schrauwen B, d’Haene M, Stroobandt D (2007) An experimental unification of reservoir computing methods. Neural Netw 20(3):391–403
Rodan A, Tino P (2011) Minimum complexity echo state network. IEEE Trans Neural Netw 22(1):131–144
Gensler A, Henze J, Sick B, Raabe N (2016) Deep learning for solar power forecasting—an approach using AutoEncoder and LSTM neural networks. In: 2016 IEEE international conference on systems, man, and cybernetics (SMC). IEEE, pp 002858–002865
Łukasik S, Kowalski PA (2014) Fully informed swarm optimization algorithms: basic concepts, variants and experimental evaluation. In: 2014 Federated conference on computer science and information systems (FedCSIS). IEEE, pp 155–61
Wang S, Phillips P, Yang J, Sun P, Zhang Y (2016) Magnetic resonance brain classification by a novel binary particle swarm optimization with mutation and time-varying acceleration coefficients. Biomed Eng (Biomedizinische Technik) 61(4):431–441
Zhang Y, Wang S, Phillips P, Dong Z, Ji G, Yang J (2015) Detection of Alzheimer’s disease and mild cognitive impairment based on structural volumetric MR images using 3D-DWT and WTA-KSVM trained by PSOTVAC. Biomed Signal Process Control 21:58–73
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Appendix: Comparison against classic GA and PSO
Appendix: Comparison against classic GA and PSO
The present appendix reports a brief comparison against classic global optimization techniques, in place of the CFSO, for the initialization of the search space to be investigated by the LM. In particular, the genetic algorithm (GA) and the classic PSO were used. Both the PSO and the GA were used considering a set of 10 agents (i.e., 10 particles for the PSO, 10 individuals for the GA). For the GA, an intermediate crossover rule was used with a 10% proportional mutation probability. No elitism was implemented. The PSO was directly derived from the FSO by forcing to 0 the H terms. Analogously to the tests described in Sect. 6.2, 300 iterations were considered: 200 for the global optimization algorithm (PSO or GA) and 100 for the refinement via LM. The test was done only on the optimal sizing parameters of the NARX; thus, N = 7, DU = 3 and DX = 1. For both algorithms GA + LM and PSO + LM, the test was repeated 150 times. Average RMSE for the GA + LM is 0.05981, and average RMSE for the PSO + LM is 0.07005. Both the methods outperform random initialization of the LM, with a distinct advantage of the GA over the PSO. Still, the PSO used is the simplest version ever formulated, and several variations [40,41,42] exist that can give much better optimization capabilities. Results are summarized in Table 5, and a time response of the system is shown in Fig. 6. As can be seen from the figure, GA and CFSO have very similar performance, and PSO shows a small accuracy degradation.
Time response for the double-tank dynamic system (bottom, black), the optimally trained neural network by CFSO + LM (bottom, red), the optimally trained neural network by GA + LM (bottom, blue), the optimally trained neural network by PSO + LM (bottom, magenta), given a time-variable input (top, pale blue). Simulations are performed on the validation set (color figure online)
Rights and permissions
About this article
Cite this article
Lozito, G.M., Salvini, A. Swarm intelligence based approach for efficient training of regressive neural networks. Neural Comput & Applic 32, 10693–10704 (2020). https://doi.org/10.1007/s00521-019-04606-x
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-019-04606-x