Skip to main content

Abstract

In constructing a classifier ensemble diversity is more important as the accuracy of its elements. To reach a diverse ensemble, one approach is to produce a pool of classifiers. Then we define a metric to evaluate the diversity value in a set of classifiers. We extract a subset of classifiers out of the pool in such a way that has a high diversity value. Usage of Bagging and Boosting as the sources of generators of diversity is another alternative. The third alternative is to partition classifiers and then select a classifier from each partition. Because of high similarity between classifiers of each partition, there is no need to let more than exactly one classifier from each of partition participate in the final ensemble. In this article, the performance of proposed framework is evaluated on some real datasets of UCI repository. Achieved results show effectiveness of the algorithm compare to the original bagging and boosting algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Blake, C.L., Merz, C.J.: UCI Repository of machine learning databases (1998), http://www.ics.uci.edu/~mlearn/MLRepository.html

  2. Breiman, L.: Bagging Predictors. Journal of Machine Learning 24(2), 123–140 (1996)

    MATH  Google Scholar 

  3. Breiman, L.: Random Forests. Machine Learning 45(1), 5–32 (2001)

    Article  MATH  Google Scholar 

  4. Freund, Y., Schapire, R.E.: A Decision-Theoretic Generalization of On-Line Learning and an Application to Boosting. J. Comput. Syst. Sci. 55(1), 119–139 (1997)

    Article  MathSciNet  MATH  Google Scholar 

  5. Giacinto, G., Roli, F.: An approach to the automatic design of multiple classifier systems. Pattern Recognition Letters 22, 25–33 (2001)

    Article  MATH  Google Scholar 

  6. Gunter, S., Bunke, H.: Creation of classifier ensembles for handwritten word recognition using feature selection algorithms. In: IWFHR (2002)

    Google Scholar 

  7. Haykin, S.: Neural Networks, a comprehensive foundation, 2nd edn. Prentice Hall International, Inc., Englewood Cliffs (1999) ISBN: 0-13-908385-5

    MATH  Google Scholar 

  8. Kuncheva, L.I.: Combining Pattern Classifiers, Methods and Algorithms. Wiley, New York (2005)

    MATH  Google Scholar 

  9. Minaei-Bidgoli, B., Topchy, A.P., Punch, W.F.: Ensembles of Partitions via Data Resampling. ITCC, 188–192 (2004)

    Google Scholar 

  10. Yang, T.: Computational Verb Decision Trees. International Journal of Computational Cognition, 34–46 (2006)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2011 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

Parvin, H., Minaei-Bidgoli, B., Beigi, A. (2011). A New Classifier Ensembles Framework. In: König, A., Dengel, A., Hinkelmann, K., Kise, K., Howlett, R.J., Jain, L.C. (eds) Knowledge-Based and Intelligent Information and Engineering Systems. KES 2011. Lecture Notes in Computer Science(), vol 6881. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-23851-2_12

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-23851-2_12

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-23850-5

  • Online ISBN: 978-3-642-23851-2

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics