Abstract
In the present study, we demonstrate the learning and recognition capabilities of our recently proposed recurrent neural network (RNN) model called stochastic continuous-time RNN (S-CTRNN). S-CTRNN can learn to predict not only the mean but also the variance of the next state of the learning targets. The network parameters consisting of weights, biases, and initial states of context neurons are optimized through maximum likelihood estimation (MLE) using the gradient descent method. First, we clarify the essential difference between the learning capabilities of conventional CTRNN and S-CTRNN by analyzing the results of a numerical experiment in which multiple fluctuating temporal patterns were used as training data, where the variance of the Gaussian noise varied among the patterns. Furthermore, we also show that the trained S-CTRNN can recognize given fluctuating patterns by inferring the initial states that can reproduce the patterns through the same MLE scheme as that used for network training.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Elman, J.L.: Finding Structure in Time. Cognitive Science 14(2), 179–211 (1990)
Pollack, J.B.: The induction of dynamical recognizers. Machine Learning 7(2-3), 227–252 (1991)
Sutskever, I., Martens, J., Hinton, G.E.: Generating Text with Recurrent Neural Networks. In: Proceedings of the 28th International Conference on Machine Learning (2011)
Kimura, M., Nakano, R.: Learning dynamical systems by recurrent neural networks from orbits. Neural Networks 11(9), 1589–1599 (1998)
Namikawa, J., Tani, J.: A model for learning to segment temporal sequences, utilizing a mixture of RNN experts together with adaptive variance. Neural Networks 21(10), 1466–1475 (2008)
Namikawa, J., Tani, J.: Building Recurrent Neural Networks to Implement Multiple Attractor Dynamics Using the Gradient Descent Method. Advances in Artificial Neural Systems 2009(14), 1–11 (2009)
Tani, J.: Model-based learning for mobile robot navigation from the dynamical systems perspective. IEEE Transactions on Systems, Man, and Cybernetics 26(3), 421–436 (1996)
Ito, M., Noda, K., Hoshino, Y., Tani, J.: Dynamic and interactive generation of object handling behaviors by a small humanoid robot using a dynamic neural network model. Neural Networks 19(3), 323–337 (2006)
Yamashita, Y., Tani, J.: Emergence of functional hierarchy in a multiple timescale neural network model: A humanoid robot experiment. PLoS Computational Biology 4(11), e1000220(2008)
Rumelhart, D.E., Hinton, G.E., Williams, R.J.: Learning Internal Representations by Error Propagation. In: Rumelhart, D.E., McClelland, D. (eds.) Parallel Distributed Processing: Explorations in the Microstructure of Cognition, pp. 318–362. MIT Press, Cambridge (1986)
Namikawa, J., Nishimoto, R., Arie, H., Tani, J.: Synthetic approach to understanding meta-level cognition of predictability in generating cooperative behavior. In: Advances in Cognitive Neurodynamics (III) Proceedings of the Third International Conference on Cognitive Neurodynamics 2011 (2013)
Murata, S., Namikawa, J., Arie, H., Sugano, S., Tani, J.: Learning to Reproduce Fluctuating Time Series by Inferring Their Time-Dependent Stochastic Properties: Application in Robot Learning Via Tutoring. IEEE Transactions on Autonomous Mental Development 5(4), 298–310 (2013)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Murata, S., Arie, H., Ogata, T., Tani, J., Sugano, S. (2014). Learning and Recognition of Multiple Fluctuating Temporal Patterns Using S-CTRNN. In: Wermter, S., et al. Artificial Neural Networks and Machine Learning – ICANN 2014. ICANN 2014. Lecture Notes in Computer Science, vol 8681. Springer, Cham. https://doi.org/10.1007/978-3-319-11179-7_2
Download citation
DOI: https://doi.org/10.1007/978-3-319-11179-7_2
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-11178-0
Online ISBN: 978-3-319-11179-7
eBook Packages: Computer ScienceComputer Science (R0)