Abstract
We have proposed an online feature extraction method called Chunk Incremental Principal Component Analysis (Chunk IPCA) where a chunk of data is trained at a time to update an eigenspace model. In this paper, we propose an extended version of Chunk IPCA in which a proper threshold for the accumulation ratio is adaptively determined such that the highest classification accuracy is maintained for a validation data set. Whenever a new chunk of training data is given, the validation set is updated in an online fashion by using the k-means clustering or through the prototype selection based on the classification results. The experimental results show that the extended version of Chunk IPCA can determine a proper threshold on an ongoing basis, resulting in keeping higher classification accuracy than the original Chunk IPCA.
The authors would like to thank Professor Shigeo Abe for his helpful comments and discussions. This research was partially supported by the Ministry of Education, Science, Sports and Culture, Grant-in-Aid for Scientific Research (C) 205002205.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Oja, E., Karhunen, J.: On Stochastic Approximation of the Eigenvectors and Eigenvalues of the Expectation of a Random Matrix. J. Math. Analysis and Application 106, 69–84 (1985)
Sanger, T.D.: Optimal Unsupervised Learning in a Single-layer Linear Feedforward Neural Network. Neural Networks 2(6), 459–473 (1989)
Chattejee, C., Kang, Z., Roychowdhury, V.: Algorithms for Accelerated Convergence of Adaptive PCA. IEEE Trans. on Neural Networks 11(2), 338–355 (2000)
Weng, J., Zhang, Y., Hwang, W.-S.: Candid Covariance-free Incremental Principal Component Analysis. IEEE Trans. on Pattern Analysis and Machine Intelligence 25(8), 1034–1040 (2003)
Pang, S., Ozawa, S., Kasabov, N.: Incremental Linear Discriminant Analysis for Classification of Data Streams. IEEE Trans. on Systems, Man, and Cybernetics, Part B 35(5), 905–914 (2005)
Hall, P., Martin, R.: Incremental Eigenanalysis for Classification. In: Proc. of British Machine Vision Conference, vol. 1, pp. 286–295 (1998)
Ozawa, S., Pang, S., Kasabov, N.: A Modified Incremental Principal Component Analysis for On-line Learning of Feature Space and Classifier. In: Zhang, C., Guesgen, H.W., Yeap, W.-K. (eds.) PRICAI 2004. LNCS (LNAI), vol. 3157, pp. 231–240. Springer, Heidelberg (2004)
Ozawa, S., Toh, S.L., Abe, S., Pang, S., Kasabov, N.: Incremental Learning of Feature Space and Classifier for Face Recognition. Neural Networks 18(5-6), 575–584 (2005)
Ozawa, S., Pang, S., Kasabov, N.: Incremental Learning of Chunk Data for On-line Pattern Classification Systems. IEEE Trans. on Neural Networks 19(6), 1061–1074 (2008)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2009 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Ozawa, S., Matsumoto, K., Pang, S., Kasabov, N. (2009). Incremental Principal Component Analysis Based on Adaptive Accumulation Ratio. In: Köppen, M., Kasabov, N., Coghill, G. (eds) Advances in Neuro-Information Processing. ICONIP 2008. Lecture Notes in Computer Science, vol 5506. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-02490-0_145
Download citation
DOI: https://doi.org/10.1007/978-3-642-02490-0_145
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-02489-4
Online ISBN: 978-3-642-02490-0
eBook Packages: Computer ScienceComputer Science (R0)