Abstract
Extreme learning machine (ELM) is a fast training scheme of single-hidden-layer feedforward neural network. How to further improve the prediction stability and accuracy of ELM in an ensemble learning way becomes one of the hot research topics in the filed of supervised learning. This paper proposes an attribute bagging-based ELM (AB-ELM) which is an ensemble learning system for classification and regression tasks by training the base ELMs on random samples of attributes instead of the entire attribute set. AB-ELM uses the sampling with replacement method to get the multiple randomized attribute subsets so as that the different data subsets can be constructed for the training of base ELMs. After obtaining a set of base ELMs, the weighted averaging method and the weighted voting method are used to generate a combination output, where the weight considers the information amount of training data subset. The relationship between the size of attribute subsets and the size of base ELMs is also discussed in AB-ELM. On 4 classification and 4 regression data sets, we verify the training and testing performances of AB-ELM in comparison with the classical ELM and the voting based ELM (V-ELM). The experimental results show that AB-ELM obtains the better prediction stability and accuracy than the classical ELM and V-ELM and thus demonstrate the effectiveness of AB-ELM.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Siuly, S., Zhang, Y.: Medical big data: neurological diseases diagnosis through medical data analysis. Data Sci. Eng. 1(2), 54–64 (2016)
Wang, J., Chen, X.: Efficient and secure storage for outsourced data: a survey. Data Sci. Eng. 1(3), 178–188 (2016)
Vargas-Solar, G., Zechinelli-Martini, J.L., Espinosa-Oviedo, J.A.: Big data management: what to keep from the past to face future challenges? Data Sci. Eng. 2(4), 328–345 (2017)
Zhai, Y., Ong, Y.S., Tsang, I.W.: The emerging “big dimensionality”. IEEE Comput. Intell. Mag. 9(3), 14–26 (2014)
Deng, W.Y., Bai, Z., Huang, G.B., et al.: A fast SVD-hidden-nodes based extreme learning machine for large-scale data analytics. Neural Netw. 77, 14–28 (2016)
Huang, G.B., Zhu, Q.Y., Siew, C.K.: Extreme learning machine: a new learning scheme of feedforward neural networks. In: IEEE International Joint Conference on Neural Networks, vol. 2, pp. 985–990 (2004)
Huang, G.B., Zhu, Q.Y., Siew, C.K.: Extreme learning machine: theory and applications. Neurocomputing 70(1–3), 489–501 (2006)
Deng, J., Frühholz, S., Zhang, Z., et al.: Recognizing emotions from whispered speech based on acoustic feature transfer learning. IEEE Access 5, 5235–5246 (2017)
Huang, Z., Yu, Y., Gu, J., et al.: An efficient method for traffic sign recognition based on extreme learning machine. IEEE Trans. Cybern. 47(4), 920–933 (2016)
Huang, G.B., Wang, D.H., Lan, Y.: Extreme learning machines: a survey. Int. J. Mach. Learn. Cybern. 2(2), 107–122 (2011)
Liu, N., Wang, H.: Ensemble based extreme learning machine. IEEE Signal Process. Lett. 17(8), 754–757 (2010)
Liang, N.Y., Huang, G.B., Saratchandran, P., et al.: A fast and accurate online sequential learning algorithm for feedforward networks. IEEE Trans. Neural Netw. 17(6), 1411–1423 (2006)
Lan, Y., Soh, Y.C., Huang, G.B.: Ensemble of online sequential extreme learning machine. Neurocomputing 72(13–15), 3391–3395 (2009)
Cao, J., Lin, Z., Huang, G.B., et al.: Voting based extreme learning machine. Inf. Sci. 185(1), 66–77 (2012)
Yang, Y., Wu, Q.M.J., Wang, Y.: Autoencoder with invertible functions for dimension reduction and image reconstruction. IEEE Trans. Syst. Man Cybern.: Syst. 48(7), 1065–1079 (2016)
Kasun, L.L.C., Yang, Y., Huang, G.B., et al.: Dimension reduction with extreme learning machine. IEEE Trans. Image Process. 25(8), 3906–3918 (2016)
Huang, G.B., Zhou, H., Ding, X., et al.: Extreme learning machine for regression and multiclass classification. IEEE Trans. Syst. Man Cybern. Part B (Cybern.) 42(2), 513–529 (2011)
Bryll, R., Gutierrez-Osuna, R., Quek, F.: Attribute bagging: improving accuracy of classifier ensembles by using random feature subsets. Pattern Recogn. 36(6), 1291–1302 (2003)
Zong, W., Huang, G.B., Chen, Y.: Weighted extreme learning machine for imbalance learning. Neurocomputing 101, 229–242 (2013)
Li, K., Kong, X., Lu, Z., et al.: Boosting weighted ELM for imbalanced learning. Neurocomputing 128, 15–21 (2014)
Dua, D., Gra, C.: UCI machine learning repository. University of California, School of Information and Computer Science, Irvine (2019). http://archive.ics.uci.edu/ml
Van Heeswijk, M., Miche, Y., Oja, E., et al.: GPU-accelerated and parallelized ELM ensembles for large-scale regression. Neurocomputing 74(16), 2430–2437 (2011)
He, Q., Shang, T., Zhuang, F., et al.: Parallel extreme learning machine for regression based on MapReduce. Neurocomputing 102, 52–58 (2013)
Acknowledgement
This paper was supported by Scientific Research Foundation of Shenzhen University for Newly-introduced Teachers (2018060), National Natural Science Foundation of China (61972261), and Open Foundation of Key Laboratory of Impression Evidence Examination and Identification Technology, Ministry of Public Security, People’s Republic of China (HJKF201901).
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2020 Springer Nature Switzerland AG
About this paper
Cite this paper
Ye, X., He, Y., Huang, J.Z. (2020). Attribute Bagging-Based Extreme Learning Machine. In: Qiu, M. (eds) Algorithms and Architectures for Parallel Processing. ICA3PP 2020. Lecture Notes in Computer Science(), vol 12453. Springer, Cham. https://doi.org/10.1007/978-3-030-60239-0_34
Download citation
DOI: https://doi.org/10.1007/978-3-030-60239-0_34
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-030-60238-3
Online ISBN: 978-3-030-60239-0
eBook Packages: Mathematics and StatisticsMathematics and Statistics (R0)