Skip to main content

Selection of the Best Base Classifier in One-Versus-One Using Data Complexity Measures

  • Conference paper
  • First Online:

Part of the book series: Lecture Notes in Computer Science ((LNAI,volume 9868))

Abstract

When dealing with multiclass problems, the most used approach is the one based on multiple binary classifiers. This approach consists of employing class binarization techniques which transforms the multiclass problem into a series of binary problems which are solved individually. Then, the resultant predictions are combined to obtain a final solution. A question arises: should the same classification algorithm be used on all binary subproblems? Or should each subproblem be tuned independently? This paper proposes a method to select a different classifier in each binary subproblem—following the one-versus-one strategy—based on the analysis of the theoretical complexity of each subproblem. The experimental results on 12 real world datasets corroborate the adequacy of the proposal when the subproblems have different structure.

This is a preview of subscription content, log in via an institution.

Buying options

Chapter
USD   29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD   39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD   54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Learn about institutional subscriptions

References

  1. Aha, D.W., Kibler, D., Albert, M.K.: Instance-based learning algorithms. Mach. Learn. 6(1), 37–66 (1991)

    Google Scholar 

  2. Bache, K., Linchman, M.: UCI machine learning repository. Univerity of California, Irvine, School of Information and Computer Sciences. http://archive.ics.uci.edu/ml/

  3. Bolón-Canedo, V., Sánchez-Maroño, N., Alonso-Betanzos, A.: Feature selection and classification in multiple class datasets: an application to KDD cup 99 dataset. Expert Syst. Appl. 38(5), 5947–5957 (2011)

    Article  Google Scholar 

  4. Cano, J.-R.: Analysis of data complexity measures for classification. Expert Syst. Appl. 40(12), 4820–4831 (2013)

    Article  MathSciNet  Google Scholar 

  5. Forman, G.: An extensive empirical study of feature selection metrics for text classification. J. Mach. Learn. Res. 3, 1289–1305 (2003)

    MATH  Google Scholar 

  6. Fürnkranz, J.: Round robin classification. J. Mach. Learn. Res. 2, 721–747 (2002)

    MathSciNet  MATH  Google Scholar 

  7. Galar, M., Fernández, A., Barrenechea, E., Bustince, H., Herrera, F.: An overview of ensemble methods for binary classifiers in multi-class problems: experimental study on one-vs-one and one-vs-all schemes. Pattern Recogn. 44(8), 1761–1776 (2011)

    Article  Google Scholar 

  8. Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I.H.: The weka data mining software: an update. ACM SIGKDD Explor. Newsl. 11(1), 10–18 (2009)

    Article  Google Scholar 

  9. Ho, T.K., Basu, M., Law, M.H.C.: Measures of geometrical complexity in classification problems. In: Basu, M., Ho, T.K. (eds.) Data Complexity in Pattern Recognition, pp. 1–23. Springer, London (2006)

    Chapter  Google Scholar 

  10. Ho, T.K., Bernadó-Mansilla, E.: Classifier domains of competence in data complexity space. In: Basu, M., Ho, T.K. (eds.) Data Complexity in Pattern Recognition, pp. 135–152. Springer, London (2006)

    Google Scholar 

  11. Kang, S., Cho, S.: Optimal construction of one-against-one classifier based on meta-learning. Neurocomputing 167, 459–466 (2015)

    Article  Google Scholar 

  12. Lorena, A.C., De Carvalho, A.C.: Evolutionary tuning of svm parameter values in multiclass problems. Neurocomputing 71(16), 3326–3334 (2008)

    Article  Google Scholar 

  13. Mendialdua, I., Martínez-Otzeta, J.M., Rodríguez-Rodríguez, I., Ruiz-Vázquez, T., Sierra, B.: Dynamic selection of the best base classifier in one versus one. Knowl.-Based Syst. 85, 298–306 (2015)

    Article  Google Scholar 

  14. Reid, S.R.: Model combination in multiclass classification. University of Colorado at Boulder (2010)

    Google Scholar 

  15. Statnikov, A., Aliferis, C., Tsardinos, I.: Gems: gene expression model selector. http://www.gems-system.org/

  16. Szepannek, G., Bischl, B., Weihs, C.: On the combination of locally optimal pairwise classifiers. Eng. Appl. Artif. Intell. 22(1), 79–85 (2009)

    Article  Google Scholar 

  17. Arizona State University. Feature selection datasets. http://featureselection.asu.edu/datasets.php

  18. Vapnik, V.N., Vapnik, V.: Statistical Learning Theory, vol. 1. Wiley, New York (1998)

    MATH  Google Scholar 

Download references

Acknowledgments

This research has been financially supported in part by the Spanish Ministerio de Economía y Competitividad (research project TIN2015-65069-C2-1-R), by European Union FEDER funds and by the Consellería de Industria of the Xunta de Galicia (research project GRC2014/035).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Laura Morán-Fernández .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2016 Springer International Publishing Switzerland

About this paper

Cite this paper

Morán-Fernández, L., Bolón-Canedo, V., Alonso-Betanzos, A. (2016). Selection of the Best Base Classifier in One-Versus-One Using Data Complexity Measures. In: Luaces , O., et al. Advances in Artificial Intelligence. CAEPIA 2016. Lecture Notes in Computer Science(), vol 9868. Springer, Cham. https://doi.org/10.1007/978-3-319-44636-3_11

Download citation

  • DOI: https://doi.org/10.1007/978-3-319-44636-3_11

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-319-44635-6

  • Online ISBN: 978-3-319-44636-3

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics