Abstract
Extreme learning machines (ELMs) are shown to be efficient and effective learning algorithms for regression and classification tasks. ELMs, however, are typically utilized to solve supervised learning problems. Only a handful of research on ELMs focuses on exploring unlabeled data. One representative work is the unsupervised extreme learning machine (US-ELM), where the standard ELM is expanded for unsupervised learning based on Laplacian regularization. However, Laplacian regularization has poor extrapolation power since it tends to bias the solution towards a constant function. In this paper, we propose a new framework termed Hessian unsupervised ELM (HUS-ELM) to enhance the unsupervised learning of ELM. In particular, Hessian regularization can properly exploit the intrinsic local geometry of the data manifold compared to Laplacian regularization. This leverages the performance of HUS-ELM in unsupervised learning problems since the Hessian regularization can correctly reflect the positional relationship between the unlabeled samples. Six publicly available datasets are used to evaluate the proposed algorithm. The experimental results indicate that the proposed method performs better than other unsupervised learning methods in terms of clustering accuracy.


Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data availability
No datasets were generated during the current study. All datasets used in this work are publicly available.
References
Hagan MT, Menhaj MB (1994) Training feedforward networks with the marquardt algorithm. IEEE Trans Neural Network 5(6):989–993
Rumelhart DE, Hinton GE, Williams RJ (1986) Learning representations by back-propagating errors. Nature 323(6088):533–536
Zhang L, Suganthan PN (2016) A comprehensive evaluation of random vector functional link networks. Information sciences 367:1094–1105
Zhang P-B, Yang Z-X (2020) A new learning paradigm for random vector functional-link network: Rvfl+. Neural Networks 122:94–105
Guo P, Lyu MR (2004) A pseudoinverse learning algorithm for feedforward neural networks with stacked generalization applications to software reliability growth data. Neurocomputing 56:101–121
Huang G-B, Zhu Q-Y, Siew C-K (2004) Extreme learning machine: a new learning scheme of feedforward neural networks. In: 2004 IEEE International Joint Conference on Neural Networks (IEEE Cat. No.04CH37541), vol. 2, pp. 985–9902. https://doi.org/10.1109/IJCNN.2004.1380068
Huang G-B, Zhou H, Ding X, Zhang R (2012) Extreme learning machine for regression and multiclass classification. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics) 42(2), 513–529. https://doi.org/10.1109/TSMCB.2011.2168604
Bartlett P. (1998) Thesamplecomplexityofp atternclassification withneuralnetworks: Thesizeoftheweightsismo reimportantthan thesizeofthenetwork. IEEETrans. Inf. Theory 44(2)
Al-Dhief FT, Baki MM, Latiff NMA, Malik NNNA, Salim NS, Albader MAA, Mahyuddin NM, Mohammed MA (2021) Voice pathology detection and classification by adopting online sequential extreme learning machine. IEEE Access 9:77293–77306. https://doi.org/10.1109/ACCESS.2021.3082565
Jamei M, Ahmadianfar I, Karbasi M, Malik A, Kisi O, Yaseen ZM (2023) Development of wavelet-based kalman online sequential extreme learning machine optimized with boruta-random forest for drought index forecasting. Eng Appl Artificial Intellig 117:105545
Zhou X, Huang J, Lu F, Zhou W, Liu P (2023) A novel compound fault-tolerant method based on online sequential extreme learning machine with cycle reservoir for turbofan engine direct thrust control. Aeros Sci Tech 132:108059
Mao W, Wang J, Xue Z (2017) An elm-based model with sparse-weighting strategy for sequential data imbalance problem. Int J Machine Learn Cyber 8:1333–1345
Yu H, Yang X, Zheng S, Sun C (2019) Active learning from imbalanced data: A solution of online weighted extreme learning machine. IEEE Trans Neural Networks Learn Syst 30(4):1088–1103. https://doi.org/10.1109/TNNLS.2018.2855446
Bhatia A, Chug A, Prakash Singh A (2020) Application of extreme learning machine in plant disease prediction for highly imbalanced dataset. J Statistics Manag Syst 23(6):1059–1068
He Q, Jin X, Du C, Zhuang F, Shi Z (2014) Clustering in extreme learning machine feature space. Neurocomputing 128:88–95
Liu T, Lekamalage CKL, Huang G-B, Lin Z (2018) Extreme learning machine for joint embedding and clustering. Neurocomputing 277:78–88
Chen J, Zeng Y, Li Y, Huang G-B (2020) Unsupervised feature selection based extreme learning machine for clustering. Neurocomputing 386:198–207
Hsu Y-S, Lin S-J (2016) An emerging hybrid mechanism for information disclosure forecasting. Int J Machine Learn Cyber 7:943–952
Baig MM, Awais MM, El-Alfy E-SM (2017) Adaboost-based artificial neural network learning. Neurocomputing 248:120–126
Huang G, Song S, Gupta JND, Wu C (2014) Semi-supervised and unsupervised extreme learning machines. IEEE Trans Cyber 44(12):2405–2417. https://doi.org/10.1109/TCYB.2014.2307349
Yang L, Yang S, Li S, Liu Z, Jiao L (2017) Incremental laplacian regularization extreme learning machine for online learning. Appl Soft Compu 59:546–555
Liu M, Liu B, Zhang C, Wang W, Sun W (2017) Semi-supervised low rank kernel learning algorithm via extreme learning machine. Int J Machine Learn Cyber 8:1039–1052
Yao L, Ge Z (2017) Deep learning of semisupervised process data with hierarchical extreme learning machine and soft sensor application. IEEE Trans Indust Electronics 65(2):1490–1498
Gharebaghi F, Amiri A (2022) Lp-mltsvm: Laplacian multi-label twin support vector machine for semi-supervised classification. IEEE Access 10:13738–13752. https://doi.org/10.1109/ACCESS.2021.3139929
Gao S, Tsang IW-H, Chia L-T (2013) Laplacian sparse coding, hypergraph laplacian sparse coding, and applications. IEEE Trans Pattern Analysis Machine Intellig 35(1):92–104. https://doi.org/10.1109/TPAMI.2012.63
Krishnasamy G, Paramesran R (2019) Multiview laplacian semisupervised feature selection by leveraging shared knowledge among multiple tasks. Signal Processing: Image Communication 70:68–78. https://doi.org/10.1016/j.image.2018.09.008
Kim K, Steinke F, Hein M (2009) Semi-supervised regression using hessian energy with an application to semi-supervised dimensionality reduction. In: Bengio, Y., Schuurmans, D., Lafferty, J., Williams, C., Culotta, A. (eds.) Advances in Neural Information Processing Systems, vol. 22. Curran Associates, Inc., https://proceedings.neurips.cc/paper/2009/file/f4552671f8909587cf485ea990207f3b-Paper.pdf
Krishnasamy G, Paramesran R (2016) Hessian semi-supervised extreme learning machine. Neurocomputing 207:560–567. https://doi.org/10.1016/j.neucom.2016.05.039
Donoho DL, Grimes C (2003) Hessian eigenmaps: Locally linear embedding techniques for high-dimensional data. Proceedings of the National Academy of Sciences 100(10), 5591–5596. https://www.pnas.org/doi/pdf/10.1073/pnas.1031596100. https://doi.org/10.1073/pnas.1031596100
Liu W, Ma T, Tao D, You J (2016) Hsae: A hessian regularized sparse auto-encoders. Neurocomputing 187:59–65. https://doi.org/10.1016/j.neucom.2015.07.119
Pan Y, Zhou Y, Liu W, Nie L (2019) Principal component analysis on graph-hessian. In: 2019 IEEE Symposium Series on Computational Intelligence (SSCI), pp. 1494–1501. https://doi.org/10.1109/SSCI44817.2019.9002887
Huang G-B, Chen L, Siew CK et al (2006) Universal approximation using incremental constructive feedforward networks with random hidden nodes. IEEE Trans. Neural Networks 17(4):879–892
Huang G-B, Chen L (2007) Convex incremental extreme learning machine. Neurocomputing 70(16–18):3056–3062
Eells J, Lemaire L (1983) Selected Topics in Harmonic Maps vol. 50. American Mathematical Soc
Belkin M, Niyogi P (2003) Laplacian eigenmaps for dimensionality reduction and data representation. Neural Computation 15(6):1373–1396
Lutkepohl H (1997) Handbook of matrices. Computa Statistics Data analysis 2(25):243
Hartigan JA, Wong MA et al (1979) A k-means clustering algorithm. Applied statistics 28(1):100–108
Ng A, Jordan M, Weiss Y (2001) On spectral clustering: Analysis and an algorithm. Advances in neural information processing systems 14
Elhamifar E, Vidal R (2013) Sparse subspace clustering: Algorithm, theory, and applications. IEEE Transactions on Pattern Analysis and Machine Intelligence 35(11):2765–2781
Ding S, Zhang N, Zhang J, Xu X, Shi Z (2017) Unsupervised extreme learning machine with representational features. Int J Machine Learn Cyber 8(2):587–595
Papadimitriou CH, Steiglitz K (1998) Combinatorial Optimization: Algorithms and Complexity. Courier Corporation
Yang Y, Shen H, Nie F, Ji R, Zhou X (2011) Nonnegative spectral clustering with discriminative regularization. Proceedings of the AAAI Conference on Artificial Intelligence 25:555–560
Acknowledgements
This work was supported by the Fundamental Research Grant Scheme (FRGS) under the Ministry of Higher Education (MOHE) with project number FRGS/1/2020/ICT02/MUSM/03/6.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Dass, S.D.S., Krishnasamy, G., Paramesran, R. et al. Hessian unsupervised extreme learning machine. Int. J. Mach. Learn. & Cyber. 15, 2013–2022 (2024). https://doi.org/10.1007/s13042-023-02012-3
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s13042-023-02012-3