Skip to main content

Studying the Hybridization of Artificial Neural Networks in HECIC

  • Conference paper
Advances in Computational Intelligence (IWANN 2011)

Part of the book series: Lecture Notes in Computer Science ((LNTCS,volume 6692))

Included in the following conference series:

  • 1828 Accesses

Abstract

One of the most relevant tasks concerning Machine Learning is the induction of classifiers, which can be used to classify or to predict. Those classifiers can be used in an isolated way, or can be combined to build a multiple classifier system. Building many-layered systems or knowing relation between different base classifiers are of special interest. Thus, in this paper we will use the HECIC system which consists of two layers: the first layer is a multiple classifier system that processes all the examples and tries to classify them; the second layer is an individual classifier that learns using the examples that are not unanimously classified by the first layer (incorporating new information). While using this system in a previous work we detected that some combinations that hybridize artificial neural networks (ANN) in one of the two layers seemed to get high-accuracy results. Thus, in this paper we have focused on the study of the improvement achieved by using different kinds of ANN in this two-layered system.

This work has been partially supported by the SESAAME project number TIN2008-06582-C03-03 of the MEC, Spain.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 84.99
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 109.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

Preview

Unable to display preview. Download preview PDF.

Unable to display preview. Download preview PDF.

References

  1. Kononenko, I., Kukar, M.: Machine Learning and Data Mining: Introduction to Principles and Algorithms. Horwood Publishing (2007)

    Google Scholar 

  2. Dietterich, T.G.: Ensemble methods in machine learning. In: Kittler, J., Roli, F. (eds.) MCS 2000. LNCS, vol. 1857, pp. 1–15. Springer, Heidelberg (2000)

    Chapter  Google Scholar 

  3. Hansen, L.K., Salamon, P.: Neural network ensembles. IEEE Transactions on Pattern Analysis and Machine Intelligence 12, 993–1001 (1990)

    Article  Google Scholar 

  4. Gama, J., Brazdil, P.: Cascade generalization. Machine Learning 41, 315–343 (2000)

    Article  MATH  Google Scholar 

  5. Utgoff, P.E., Stracuzzi, D.J.: Many-layered learning. Neural Computation 14(10), 2497–2529 (2002)

    Article  MATH  Google Scholar 

  6. Chindaro, S., Sirlantzis, K., Fairhurst, M.C.: Modelling multiple-classifier relationships using bayesian belief networks. In: Haindl, M., Kittler, J., Roli, F. (eds.) MCS 2007. LNCS, vol. 4472, pp. 312–321. Springer, Heidelberg (2007)

    Chapter  Google Scholar 

  7. Ramos-Jiménez, G., del Campo-Ávila, J., Morales-Bueno, R.: Hybridizing ensemble classifiers with individual classifiers. In: International Conference on Intelligent Systems Design and Applications. Workshop on Hybrid Learning for Artificial Neural Networks: Architectures and Applications, pp. 199–202. IEEE Computer Society, Los Alamitos (2009)

    Google Scholar 

  8. Ramos-Jiménez, G., del Campo-Ávila, J., Morales-Bueno, R.: ML-CIDIM: Multiple layers of multiple classifier systems based on CIDIM. In: Ślęzak, D., Yao, J., Peters, J.F., Ziarko, W.P., Hu, X. (eds.) RSFDGrC 2005. LNCS (LNAI), vol. 3642, pp. 138–146. Springer, Heidelberg (2005)

    Chapter  Google Scholar 

  9. Freund, Y.: Boosting a weak learning algorithm by majority. Information and Computation 121, 256–285 (1995)

    Article  MathSciNet  MATH  Google Scholar 

  10. Breiman, L.: Bagging predictors. Machine Learning 24(2), 123–140 (1996)

    MATH  Google Scholar 

  11. Aslam, J.A., Decatur, S.E.: General bounds on statistical query learning and pac learning with noise via hypothesis boosting. Information and Computation 141, 85–118 (1998)

    Article  MathSciNet  MATH  Google Scholar 

  12. Bauer, E., Kohavi, R.: An empirical comparison of voting classification algorithms: bagging, boosting, and variants. Machine Learning 36, 105–142 (1999)

    Article  Google Scholar 

  13. Freund, Y., Schapire, R.E.: Experiments with a new boosting algorithm. In: Proceedings of the 13th International Conference on Machine Learning (ICML-1996), pp. 146–148 (1996)

    Google Scholar 

  14. Aha, D.W., Kibler, D., Albert, M.K.: Instance-based learning algorithms. Machine Learning 6, 37–66 (1991)

    Google Scholar 

  15. Quinlan, J.R.: C4.5: Programs for machine learning. Morgan Kaufmann, San Francisco (1993)

    Google Scholar 

  16. John, G.H., Langley, P.: Estimating continuous distributions in bayesian classifiers. In: Proceedings of the Eleventh Annual Conference on Uncertainty in Artificial Intelligence (UAI-1995), pp. 338–345. Morgan Kaufmann, San Francisco (1995)

    Google Scholar 

  17. Gill, P.E., Murray, W., Wright, M.H.: Practical optimization. Academic Press, London (1981)

    MATH  Google Scholar 

  18. Kohonen, T.: Self-organizing maps. Springer-Verlag, New York, Inc. (1997)

    Google Scholar 

  19. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I.H.: The WEKA data mining software: An update. In: SIGKDD Explorations, vol. 11, pp. 10–18 (2009)

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2011 Springer-Verlag Berlin Heidelberg

About this paper

Cite this paper

del Campo-Ávila, J., Ramos-Jiménez, G., Pérez-García, J., Morales-Bueno, R. (2011). Studying the Hybridization of Artificial Neural Networks in HECIC. In: Cabestany, J., Rojas, I., Joya, G. (eds) Advances in Computational Intelligence. IWANN 2011. Lecture Notes in Computer Science, vol 6692. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-21498-1_18

Download citation

  • DOI: https://doi.org/10.1007/978-3-642-21498-1_18

  • Publisher Name: Springer, Berlin, Heidelberg

  • Print ISBN: 978-3-642-21497-4

  • Online ISBN: 978-3-642-21498-1

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics