Abstract
When performing predictive modeling, ensembles are often utilized in order to boost accuracy. The problem of how to maximize ensemble accuracy is, however, far from solved. In particular, the relationship between ensemble diversity and accuracy is, especially for classification, not completely understood. More specifically, the fact that ensemble diversity and base classifier accuracy are highly correlated, makes it necessary to balance these properties instead of just maximizing diversity. In this study, three standard techniques to obtain implicit diversity in neural network ensembles are evaluated using 14 UCI data sets. The experiments show that standard resampling; i.e. dividing the training data by instances, produces more diverse models, but at the expense of base classifier accuracy, thus resulting in less accurate ensembles. Building ensembles using neural networks with heterogeneous architectures improves test set accuracies, but without actually increasing the diversity. The results regarding resampling using features are inconclusive, the ensembles become more diverse, but the level of test set accuracies is unchanged. For the setups evaluated, ensemble training accuracy and base classifier training accuracy are positively correlated with ensemble test accuracy, but the opposite holds for diversity; i.e. ensembles with low diversity are generally more accurate.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Dietterich, T.G.: Machine learning research: four current directions. The AI Magazine 18, 97–136 (1997)
Krogh, A., Vedelsby, J.: Neural network ensembles, cross validation, and active learning. In: Advances in Neural Information Processing Systems, vol. 2, pp. 650–659. Morgan Kaufmann, San Mateo (1995)
Brown, G., Wyatt, J., Harris, R., Yao, X.: Diversity Creation Methods: A survey and Categorisation. Journal of Information Fusion 6(1), 5–20 (2005)
Sharkey, N., Neary, J., Sharkey, A.: Searching Weight Space for Backpropagation Solution Types. In: Current trends in Connectionism: Proceedings of the 1995 Swedish Conference on Connectionism, pp. 103–120 (1995)
Duin, R.P.W., Tax, D.M.J.: Experiments with classifier combining rules. In: Kittler, J., Roli, F. (eds.) MCS 2000. LNCS, vol. 1857, pp. 30–44. Springer, Heidelberg (2000)
Kuncheva, L.I., Whitaker, C.J.: Measures of Diversity in Classifier Ensembles and Their Relationship with the Ensemble Accuracy. Machine Learning 51, 181–207 (2003)
Johansson, U., Löfström, T., Niklasson, L.: The Importance of Diversity in Neural Network Ensembles - An Empirical Investigation. In: The International Joint Conference on Neural Networks, pp. 661–666. IEEE Press, Orlando (2007)
Blake, C.L., Merz, C.J., Repository, U.C.I.: of machine learning databases, University of California, Department of Information and Computer Science (1998)
Author information
Authors and Affiliations
Editor information
Rights and permissions
Copyright information
© 2008 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Johansson, U., Löfström, T., Niklasson, L. (2008). Evaluating Standard Techniques for Implicit Diversity. In: Washio, T., Suzuki, E., Ting, K.M., Inokuchi, A. (eds) Advances in Knowledge Discovery and Data Mining. PAKDD 2008. Lecture Notes in Computer Science(), vol 5012. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-68125-0_54
Download citation
DOI: https://doi.org/10.1007/978-3-540-68125-0_54
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-68124-3
Online ISBN: 978-3-540-68125-0
eBook Packages: Computer ScienceComputer Science (R0)