Abstract
VSF−Network is a neural network model that learns dynamical patterns. It is hybrid neural network combining a chaos neural network and a hierarchical neural network. The hierarchical neural network learns patterns and the chaos neural network monitors behavior of neurons in the hierarchical neural network. In this paper, two theoretical backgrounds of VSF−Network are introduced. An incremental learning framework using chaos neural networks is introduced. The monitoring by chaos neural network is based on clusters generated by synchronous vibration. Using the monitoring results, redundant neurons in the hierarchical neural network are found and they are used for learning of new patters. The second background is about the pattern recognition by combining learned patterns. This is explained by code words expression used in multi-level discrimination. Through an experiment, both the incremental learning capability and the pattern recognition are shown.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Inamura, T., Tanie, H., Nakamura, Y.: Proto-symbol development and manipulation in the geometry of stochastic model for motion generation and recognition. Technical Report NC2003-65, IEICE (2003)
Kadone, H., Nakamura, Y.: Symbolic memory of motion patterns using hierarchical bifurcations of attanctors in an associative memory model. J. Robot Soc. Jpn. 25, 249–258 (2007)
Chandler, D.: Semiotics for Beginners. Routledge, London (1995)
Kakemoto, Y., Nakasuka, S.: Neural assembly generation by selective connection weight updating. In: Proceedings of IjCNN 2010 (2010)
Giraud-Carrier, C.: A note on the utility of incremental learning. AI Commun. 13, 215–223 (2000)
Lin, M., Tang, K., Yao, X.: Incremental learning by negative correlation leaning. In: Proceedings of IJCNN 2008 (2008)
Hopfield, J.: Neurons with graded response have collective computational properties like those of two-stage neurons. Proc. Nat. Acad. Sci. U.S.A. 81, 13088–3092 (1984)
Aihara, T., Tanabe, T., Toyoda, M.: Chaotic neural networks. Phys. Lett. 144A, 333–340 (1990)
Kaneko, K.: Chaotic but regular posi-nega switch among coded attractor by cluster size variation. Phys. Rev. Lett. 63, 219 (1989)
Komuro, M.: A mechanism of chaotic itinerancy in globally coupled maps. In: Dynamical Systems (NDDS 2002) (2002)
LeCun, Y., Boser, B., Denker, J.S., Henderson, D., Howard, R.E., Hubbard, W., Jackel, L.D.: Backpropagation applied to handwritten zip code recognition. Neural Comput. 1, 541–551 (1989)
Sejnowski, T.J., Rosenberg, C.R.: Nettalk: a parallel network that learns to read aloud. In: Anderson, J.A., Rosenfeld, E. (eds.) Neurocomputing: Foundations of Research, pp. 661–672. MIT Press, Cambridge (1988)
Dietterich, T.G., Bakiri, G.: Solving multiclass learning problems via error-correcting output codes. J. Artif. Intell. Res. 2, 263–286 (1995)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2017 Springer International Publishing AG
About this paper
Cite this paper
Kakemoto, Y., Nakasuka, S. (2017). Learning Symbols by Neural Network. In: Angelov, P., Manolopoulos, Y., Iliadis, L., Roy, A., Vellasco, M. (eds) Advances in Big Data. INNS 2016. Advances in Intelligent Systems and Computing, vol 529. Springer, Cham. https://doi.org/10.1007/978-3-319-47898-2_10
Download citation
DOI: https://doi.org/10.1007/978-3-319-47898-2_10
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-47897-5
Online ISBN: 978-3-319-47898-2
eBook Packages: EngineeringEngineering (R0)