Abstract
Neural networks are a powerful computational architecture for modeling data, but optimizing the connection weights can be very difficult. Flexible neural trees (FNTs) are good at finding a globally near-optimal network to fit a dataset, using evolutionary algorithms and particle swarm optimization. We show that putting the two methods together can yield very good results. The FNT solution can be embedded into a larger neural network that is then optimized using backpropagation. The combination of the two methods outperforms either method alone.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Bengio, Y., Lamblin, P., Popovici, D., Larochelle, H.: Greedy layer-wise training of deep networks. In: Proceedings of the 19th International Conference on Neural Information Processing Systems (NIPS 2006), pp. 153–160. MIT Press, Cambridge (2006)
Chen, Y., Abraham, A., Yang, B.: Feature selection and classification using flexible neural tree. Neurocomputing 70(1), 305–313 (2006)
Chen, Y., Yang, B., Dong, J.: Evolving flexible neural networks using ant programming and PSO algorithm. In: Yin, F.-L., Wang, J., Guo, C. (eds.) ISNN 2004. LNCS, vol. 3173, pp. 211–216. Springer, Heidelberg (2004). doi:10.1007/978-3-540-28647-9_36
Chen, Y., Yang, B., Dong, J., Abraham, A.: Time-series forecasting using flexible neural tree model. Inf. Sci. 174(3), 219–235 (2005)
Chen, Z., Peng, L., Gao, C., Yang, B., Chen, Y., Li, J.: Flexible neural trees based early stage identification for IP traffic. Soft Comput. 21(8), 2035–2046 (2017)
Hinton, G.E.: A practical guide to training restricted boltzmann machines. In: Montavon, G., Orr, G.B., Müller, K.-R. (eds.) Neural Networks: Tricks of the Trade. LNCS, vol. 7700, pp. 599–619. Springer, Heidelberg (2012). doi:10.1007/978-3-642-35289-8_32
Le Cun, Y., Boser, B., Denker, J.S., Henderson, D., Howard, R.E., Hubbard, W., Jackel, L.D.: Handwritten digit recognition with a back-propagation network. In: NIPS 1990, pp. 396–404 (1990)
Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning representations by back-propagating errors. Nature 323(6088), 533–536 (1986)
Acknowledgments
This research was supported by the National Key Research and Development Program of China (No. 2016YFC0106000), the Youth Science and Technology Star Program of Jinan City (201406003).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Wu, P., Orchard, J. (2017). Using Flexible Neural Trees to Seed Backpropagation. In: Liu, D., Xie, S., Li, Y., Zhao, D., El-Alfy, ES. (eds) Neural Information Processing. ICONIP 2017. Lecture Notes in Computer Science(), vol 10634. Springer, Cham. https://doi.org/10.1007/978-3-319-70087-8_12
Download citation
DOI: https://doi.org/10.1007/978-3-319-70087-8_12
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-70086-1
Online ISBN: 978-3-319-70087-8
eBook Packages: Computer ScienceComputer Science (R0)