Skip to main content

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 3697))

Included in the following conference series:

  • 3516 Accesses

Abstract

In this paper we introduce a model of ensemble of linear perceptrons. The objective of the ensemble is to automatically divide the feature space into several regions and assign one ensemble member into each region and training the member to develop an expertise within the region. Utilizing the proposed ensemble model, the learning difficulty of each member can be reduced, thus achieving faster learning while guaranteeing the overall performance.

An erratum to this chapter can be found at http://dx.doi.org/10.1007/11550907_163 .

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Baxt, W.: Improving Accuracy of Artificial Neural Network using Multiple Differently Trained Networks. Neural Computation 4, 108–121 (1992)

    Article  Google Scholar 

  2. Sharkey, A.: On Combining Artificial Neural Nets. Connection Science, Vol 9(3,4), 299–313 (1996)

    Article  Google Scholar 

  3. Hashem, S.: Optimal Linear Combination of Neural Networks. Neural Networks 10(4), 559–614 (1996)

    Google Scholar 

  4. Freund, Y.: Boosting a weak learning algorithm by Majority. Information and Computation 121(2), 256–285 (1995)

    Article  MATH  MathSciNet  Google Scholar 

  5. Jacobs, R., Jordan, M., Nowlan, S., Hinton, G.: Adaptive Mixture of Local Experts. Neural Computation 3, 79–87 (1991)

    Article  Google Scholar 

  6. Minsky, M., Papert, S.: Perceptron. The MIT Press, Cambridge (1969)

    Google Scholar 

  7. Rumelhart, D.E., McClelland, J.: Learning Internal Representation by Error Propagation. In: Parallel Distributed Processing, vol. 1, pp. 318–362. MIT Press, Cambridge (1984)

    Google Scholar 

  8. Mangasarian, O.L., Wolberg, W.H.: Cancer Diagnosis via linear programming. SIAM News 23(5), 1–18 (1990)

    Google Scholar 

  9. UCI Machine Learning Repository, http://www.ics.uci.edu/~mlearn/MLRepository.html

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2005 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Hartono, P., Hashimoto, S. (2005). Learning with Ensemble of Linear Perceptrons. In: Duch, W., Kacprzyk, J., Oja, E., Zadrożny, S. (eds) Artificial Neural Networks: Formal Models and Their Applications – ICANN 2005. ICANN 2005. Lecture Notes in Computer Science, vol 3697. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11550907_19

Download citation

  • DOI: https://doi.org/10.1007/11550907_19

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-540-28755-1

  • Online ISBN: 978-3-540-28756-8

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics