Abstract
Recently, multiple classifier systems (MCS) have been used for practical applications to improve classification accuracy. Self-generating neural networks (SGNN) are one of the suitable base-classifiers for MCS because of their simple setting and fast learning. However, the computational cost of the MCS increases in proportion to the number of SGNN. In an earlier paper, we proposed a pruning method for the structure of the SGNN in the MCS to reduce the computational cost. In this paper, we propose a novel pruning method for effective processing. The pruning method is constructed from an on-line pruning method and an off-line pruning method. We implement the pruned MCS with two sampling methods. Experiments have been conducted to compare the pruned MCS with the unpruned MCS, the MCS based on C4.5, and k-nearest neighbor method. The results show that the pruned MCS can improve its classification accuracy as well as reducing the computational cost.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Han, J. and Kamber, M.: Data Mining: Concepts and Techniques. Morgan Kaufmann Publishers, San Francisco, CA (2000)
Quinlan, J. R.: Bagging, Boosting, and C4.5. Proceedings of the Thirteenth National Conference on Artificial Intelligence, AAAI Press and the MIT Press, Portland, OR (1996) 725–730
Rätsch, G., Onoda, T. and Müller, K.-R.: Soft margins for AdaBoost. Machine Learning 42(3) (2001) 287–320
Bishop, C. M.: Neural Networks for Pattern Recognition. Oxford University Press, New York (1995)
Duda, R. O., Hart, P. E. and Stork, D. G.: Pattern Classification. 2nd ed. John Wiley & Sons Inc., New York (2000).
Wen, W. X., Jennings, A. and Liu, H.: Learning a neural tree. Proceedings of the International Joint Conference on Neural Networks 2, Beijing, China (1992) 751–756
Kohonen, T.: Self-Organizing Maps. Springer-Verlag, Berlin (1995)
Inoue, H. and Narihisa, H.: Improving generalization ability of self-generating neural networks through ensemble averaging. In: Terano, T., Liu, H., and Chen, A. L. P. (eds.): The Fourth Pacific-Asia Conference on Knowledge Discovery and Data Mining, LNAI 1805, Springer-Verlag, Berlin (2000) 177–180
Inoue, H. and Narihisa, H.: Optimizing a multiple classifier system. In: Ishizuka, M. and Sattar, A. (eds.): PRICAI2002: Trends in Artificial Intelligence, LNAI 2417, Springer-Verlag, Berlin (2002) 285–294
Stone, M.: Cross-validation: A review. Math. Operationsforsch. Statist., Ser. Statistics 9(1) (1978) 127–139
Breiman, L.: Bagging predictors. Machine Learning 24 (1996) 123–140
Quinlan, J. R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo, CA (1993)
Blake, C. L. and Merz, C. J.: UCI repository of machine learning databases, University of California, Irvine, Dept of Information and Computer Science (1998) Datasets is available at http://www.ics.uci.edu/~mlearn/MLRepository.html
Patrick, E. A. and Fischer, F. P.: A generalized k-nearest neighbor rule. Information and Control 16(2) (1970) 128–152
Breiman, L., Friedman, J., Olshen, R. and Stone, C.: Classification and Regression Trees. Wadsworth, Belmont, CA (1984)
Cohen, S. and Intrator, N.: Forward and backward selection in regression hybrid network. In: Roli, F. and Kittler, J. (eds.): Multiple Classifier Systems, Third International Workshop, LNCS 2364, Springer-Verlag, Berlin (2002) 98–107
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2003 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Inoue, H., Narihisa, H. (2003). Improving Performance of a Multiple Classifier System Using Self-generating Neural Networks. In: Windeatt, T., Roli, F. (eds) Multiple Classifier Systems. MCS 2003. Lecture Notes in Computer Science, vol 2709. Springer, Berlin, Heidelberg. https://doi.org/10.1007/3-540-44938-8_26
Download citation
DOI: https://doi.org/10.1007/3-540-44938-8_26
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-40369-2
Online ISBN: 978-3-540-44938-6
eBook Packages: Springer Book Archive