Abstract
In this paper, a novel selective ensemble method based on the improved negative correlation learning is proposed. To make the proposed ensemble strategy more robust against noise, correntropy is utilized to substitute mean square error (MSE). Moreover, an L1-norm based regularization term of ensemble weights is incorporated into the objective function of the proposed ensemble strategy to fulfill the task of selective ensemble. The half-quadratic optimization technique and the surrogate function method are used to solve the optimization problem of the proposed ensemble strategy. Experimental results on two synthetic data sets and the five benchmark data sets demonstrate that the proposed method is superior to the single radial basis function neural network (RBFNN).
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Hansen, L.K., Salamon, P.S.: Neural network ensembles. IEEE Transactions on Pattern Analysis and Machine Intelligence 12(10), 993–1001 (1990)
Liu, Y., Yao, X.: Ensemble learning via negative correlation. Neural Networks 12(10), 1399–1404 (1999)
Liu, Y., Yao, X.: Simultaneous training of negatively correlated neural networks in ensemble. IEEE Transactions on Systems, Man, and Cybernetics-Part B 29(6), 716–725 (1999)
Chan, Z.S., Kasavov, N.: Fast neural network ensemble learning via negative-correlation data correction. IEEE Transactions on Neural Networks 16(6), 1707–1710 (2005)
Islam, M.M., Yao, X., Nirjon, S.M.S., Islam, M.A., Murase, K.: Bagging and Boosting negatively correlated neural networks. IEEE Transactions on Systems, Man, and Cybernetics-Part B: Cybernetics 38(3), 771–784 (2008)
Chen, H., Yao, X.: Regularized negative correlation learning for neural network ensembles. IEEE Transactions on Neural Networks 20(12), 1962–1979 (2009)
Alhamdoosh, M., Wang, D.: Fast decorrelated neural network ensembles with random weights. Information Sciences 264, 104–117 (2014)
Zhou, Z.H., Wu, J., Tang, W.: Ensembling neural networks: many could be better than all. Artificial Intelligence 137(1-2), 239–263 (2002)
Haykin, S.: Neural Networks: A Comprehensive Foundation, 2nd edn. Prentice-Hall, New York (1999)
Rumelhart, D.E., McClelland, J.L.: Parallel Distributed Processing: Exploration in the Microstructure of Cognition. MIT Press, Cambridge (1986)
Bishop, C.M.: Neural Networks for Pattern Recognition. Oxford University Press, Oxford (1995)
Chen, S., Wu, Y., Luk, B.L.: Combined genetic algorithm optimization and regularized orthogonal least squares learning for radial basis function networks. IEEE Transactions on Neural Networks 10(5), 1239–1243 (1999)
Liu, W., Pokharel, P.P., Principe, J.C.: Correntropy: properties and applications in non-Gaussian signal processing. IEEE Transactions on Signal Process 55(11), 5286–5297 (2007)
Santamaria, I., Pokharel, P.P., Principe, J.C.: Generalized correlation function: definition, properties, and application to blind equalization. IEEE Transactions on Signal Process 54(6), 2187–2197 (2006)
Vapnik, V.: The nature of statistical learning theory. Springer, New York (1995)
Yuan, X, Hu, B.G.: Robust feature extraction via information theoretic learning. In: Proceedings of The 26th Annual International Conference on Machine Learning, pp. 1193–1200. ACM, New York (2009)
Rockfellar, R.: Convex analysis. Princeton University Press, Princeton (1970)
Yang, Shuang-Hong, Zha, Hongyuan, Zhou, SKevin, Hu, Bao-Gang: Variational Graph Embedding for Globally and Locally Consistent Feature Extraction. In: Buntine, Wray, Grobelnik, Marko, Mladenić, Dunja, Shawe-Taylor, John (eds.) ECML PKDD 2009, Part II. LNCS, vol. 5782, pp. 538–553. Springer, Heidelberg (2009)
Boyd, S., Vandenberghe, L.: Convex optimization. Cambridge University Press, Cambridge (2004)
Lannge, K., Hunter, D., Yang, I.: Optimization transfer using surrogate objective functions. Journal of Computational and Graphical Statistics 9(1), 1–59 (2000)
Silverman, B.W.: Density Estimation for Statistics and Data Analysis. Chapman and Hall, London (1986)
Blake C.L., Merz C.J., UCI Repository of Machine Learning Databases, Department of Information and Computer Science, University of California, Irvine, CA (1998). http://www.ics.uci.edu/mlearn/MLRepository.html
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Xing, H., Liu, L., Li, S. (2014). Selective Ensemble of RBFNNs Based on Improved Negative Correlation Learning. In: Wang, X., Pedrycz, W., Chan, P., He, Q. (eds) Machine Learning and Cybernetics. ICMLC 2014. Communications in Computer and Information Science, vol 481. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-662-45652-1_31
Download citation
DOI: https://doi.org/10.1007/978-3-662-45652-1_31
Published:
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-662-45651-4
Online ISBN: 978-3-662-45652-1
eBook Packages: Computer ScienceComputer Science (R0)