Skip to main content
Log in

Further improvements on extreme learning machine for interval neural network

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

The interval extreme learning machine (IELM) (Yang et al. in Neural Comput Appl 27(1):3–8, 2016) is a newly proposed regression algorithm to deal with the data with interval-valued inputs and interval-valued output. In this paper, we firstly analyze the disadvantages of IELM and further point out that IELM is actually a slight variant of fuzzy regression analysis using neural networks (Ishibuchi and Tanaka in Fuzzy Sets Syst 50(3):257–265, 1992). Then, we propose a new interval-valued ELM (IVELM) model to handle the interval-valued data regression. IVELM does not require any iterative adjustment to network weights and thus has the extremely fast training speed. The experimental results on data sets used in (Yang et al. 2016) demonstrate the feasibility and effectiveness of IVELM which obtains the better predictive performance and faster learning speed than IELM.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

Notes

  1. Although the authors described the regression problem with multiple interval-valued outputs, in fact, they only used IELM to handle the regression problem with single interval-valued output. The experimental data sets used in IELM also confirmed this fact. Thus, there is only one interval-valued output which is considered in the following discussion.

References

  1. He YL, Wang XZ, Huang JZX (2016) Fuzzy nonlinear regression analysis using a random weight network. Inf Sci 364–365:222–240

    Article  Google Scholar 

  2. Hoerl AE, Kennard RW (2000) Ridge regression: biased estimation for nonorthogonal problems. Technometrics 42(1):80–86

    Article  MATH  Google Scholar 

  3. Huang GB, Zhu QY, Siew CK (2006) Extreme learning machine: theory and applications. Neurocomputing 70(1):489–501

    Article  Google Scholar 

  4. Ishibuchi H, Kwon K, Tanaka H (1995) A learning algorithm of fuzzy neural networks with triangular fuzzy weights. Fuzzy Sets Syst 71(3):277–293

    Article  Google Scholar 

  5. Ishibuchi H, Tanaka H (1992) Fuzzy regression analysis using neural networks. Fuzzy Sets Syst 50(3):257–265

    Article  MathSciNet  Google Scholar 

  6. Rumelhart DE, Hinton GE, Williams RG (1986) Learning representations by back-propagating errors. Nature 323:533–536

    Article  MATH  Google Scholar 

  7. Yang D, Li Z, Wu W (2016) Extreme learning machine for interval neural networks. Neural Comput Appl 27(1):3–8

    Article  Google Scholar 

Download references

Acknowledgements

We thank the editor-in-chief and two anonymous reviewers for their valuable comments and suggestions. This work is supported by National Natural Science Foundations of China (61503252) and China Postdoctoral Science Foundations (2015M572361 and 2016T90799).

Author information

Authors and Affiliations

Authors

Corresponding authors

Correspondence to Chong Liu or Yu-lin He.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Yang, Lf., Liu, C., Long, H. et al. Further improvements on extreme learning machine for interval neural network. Neural Comput & Applic 29, 311–318 (2018). https://doi.org/10.1007/s00521-016-2727-4

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-016-2727-4

Keywords

Navigation