Skip to main content
Log in

Model selection of extreme learning machine based on multi-objective optimization

  • Extreme Learning Machine’s Theory & Application
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

As a novel learning algorithm for single-hidden-layer feedforward neural networks, extreme learning machines (ELMs) have been a promising tool for regression and classification applications. However, it is not trivial for ELMs to find the proper number of hidden neurons due to the nonoptimal input weights and hidden biases. In this paper, a new model selection method of ELM based on multi-objective optimization is proposed to obtain compact networks with good generalization ability. First, a new leave-one-out (LOO) error bound of ELM is derived, and it can be calculated with negligible computational cost once the ELM training is finished. Furthermore, the hidden nodes are added to the network one-by-one, and at each step, a multi-objective optimization algorithm is used to select optimal input weights by minimizing this LOO bound and the norm of output weight simultaneously in order to avoid over-fitting. Experiments on five UCI regression data sets are conducted, demonstrating that the proposed algorithm can generally obtain better generalization performance with more compact network than the conventional gradient-based back-propagation method, original ELM and evolutionary ELM.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

References

  1. Huang GB, Zhu QY, Siew C (2004) Extreme learning machine: a new learning scheme of feedforward neural networks. In: Proceedings of International Joint Conference on Neural Networks (IJCNN2004), vol 2, pp 25–29

  2. Huang GB, Zhu QY, Siew C (2006) Extreme learning machine: theory and applications. Neurocomputing 70:489–501

    Article  Google Scholar 

  3. Chen FL, Ou TY (2011) Sales forecasting system based on gray extreme learning machine with taguchi method in retail industry. Expert Syst Appl 38:1336–1345

    Article  Google Scholar 

  4. Minhas R, Baradarani A, Seifzadeh S, Wu QMJ (2010) Human action recognition using extreme learning machine based on visual vocabularies. Neurocomputing 73:1906–1917

    Article  Google Scholar 

  5. Mohammed A, Wu QMJ, Sid-Ahmed M (2010) Application of wave atoms decomposition and extreme learning machine for fingerprint classification. Lect Notes Comput Sci 6112:246–256

    Article  Google Scholar 

  6. Huang GB, Wang D, Lan Y (2011) Extreme learning machines: a survey. Int J Mach Learn Cybern 2:107–122

    Article  Google Scholar 

  7. Feng GR, Huang GB, Lin QP, Gay R (2009) Error minimized extreme learning machine with growth of hidden nodes and incremental learning. IEEE Trans Neural Netw 20:1352–1357

    Article  Google Scholar 

  8. Lan Y, Soh Y, Huang GB (2010) Random search enhancement of error minimized extreme learning machine. In: Proceedings of European Symposium on Artificial Neural Networks (ESANN 2010), pp 327–332

  9. Li K, Huang GB, Ge SS (2010) Fast construction of single hidden layer feedforward networks. In: Rozenberg G, Bäck T, Kok JN (eds) Handbook of Natural Computing. Springer, Berlin

  10. Lan Y, Soh Y, Huang GB (2006) Two-stage extreme learning machine for regression. Neurocomputing 73:3028–3038

    Article  Google Scholar 

  11. Zhu QY, Qin AK, Suganthan P, Huang GB (2005) Evolutionary extreme learning machine. Pattern Recognit 38:1759–1763

    Article  MATH  Google Scholar 

  12. Huang GB, Ding XJ, Zhou H (2010) Optimization method based extreme learning machine for classification. Neurocomputing 74:155–163

    Article  Google Scholar 

  13. Cawley G, Talbot N (2007) Preventing over-fitting in model selection via bayesian regularisation of the hyper-parameters. J Mach Learn Res 8:841–861

    MATH  Google Scholar 

  14. Huang GB, Babri H (1998) Upper bounds on the number of hidden neurons in feedforward networks with arbitrary bounded nonlinear activation functions. IEEE Trans Neural Netw 9:224–229

    Article  Google Scholar 

  15. Girosi F, Jones M, Poggio T (1995) Regularization theory and neural network architectures. Neural Comput 7:219–269

    Article  Google Scholar 

  16. Evgeniou T, Pontil M, Poggio T (2000) Regularization networks and support vector machines. Adv Comput Math 13:1–50

    Article  MathSciNet  MATH  Google Scholar 

  17. Frénay B, Verleysen M (2010) Using SVMs with randomised feature spaces: an extreme learning approach. In: Proceedings of The 18th European Symposium on Artificial Neural Networks (ESANN 2010), pp 315–320

  18. Myers R (1990) Classical and modern regression with applications. Duxbury Press, CA

    Google Scholar 

  19. Kohavi R (1995) A study of cross-validation and bootstrap for accuracy estimation and model selection. In: Proceedings of the Fourteenth International Conference on Artificial Intelligence (IJCAI), pp 1137–1143

  20. Bartlett P (1998) The sample complexity of pattern classification with neural networks: the size of the weights is more important than the size of the network. IEEE Trans Inf Theory 44:525–536

    Article  MathSciNet  MATH  Google Scholar 

  21. Huang VL, Suganthan P, Liang JJ (2006) Comprehensive learning particle swarm optimizer for solving multiobjective optimization problems. Int J Intell Syst 21:209–226

    Article  MATH  Google Scholar 

  22. Zhu QY, Huang GB (2004) Source codes of ELM algorithm. In: http://www.ntu.edu.sg/home/egbhuang/, School of Electrical and Electronic Engineering, Nanyang Technological University, Singapore

  23. Newman D, Hettich S, Blake C (1998) UCI Repository of machine learning databases. In: http://www.ics.uci.edu/mlearn/MLRepository.html, Department of Information and Computer Science, University of California, Irvine, CA

  24. Wang G, Yeung D, Lochovsky F (2008) A new solution path algorithm in support vector regression. IEEE Trans Neural Netw 19:1753–1767

    Article  Google Scholar 

Download references

Acknowledgments

This work was supported by the National Natural Science Foundation of China (60873104). We also thank the author Suganthan of [21] for providing implementation of MOCLPSO.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Wentao Mao.

Rights and permissions

Reprints and permissions

About this article

Cite this article

Mao, W., Tian, M., Cao, X. et al. Model selection of extreme learning machine based on multi-objective optimization. Neural Comput & Applic 22, 521–529 (2013). https://doi.org/10.1007/s00521-011-0804-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-011-0804-2

Keywords

Navigation