Abstract
One-against-all and one-against-one are two popular methodologies for reducing multiclass classification problems into a set of binary classifications. In this paper, we are interested in the performance of both one-against-all and one-against-one for basic classification algorithms, such as decision tree, naïve bayes, support vector machine, and logistic regression. Since both one-against-all and one-against-one work like creating a classification committee, they are expected to improve the performance of classification algorithms. However, our experimental results surprisingly show that one-against-all worsens the performance of the algorithms on most datasets. One-against-one helps, but performs worse than the same iterations of bagging these algorithms. Thus, we conclude that both one-against-all and one-against-one should not be used for the algorithms that can perform multiclass classifications directly. Bagging is an better approach for improving their performance.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Mitchell, T.M.: Machine Learning, pp. 52–77. McGraw-Hill (1997)
Rifkin, R., Klautau, A.: In Defense of One-vs-All Classification. Journal of Machine Learning Research 5, 101–141 (2004)
Tsujinishi, D., Koshiba, Y., Abe, S.: Why Pairwise is Better Than One-Against-All or All-at-Once. In: IEEE International Joint Conference, pp. 693–698 (2007)
Frank, A., Asuncion, A.: UCI Machine Learning Repository. University of California, School of Information and Computer Science (2010), http://archive.ica.uci.edu/ml/datasets.html?
Hsu, C.W., Lin, C.W.: A Comparison of Methods for Multiclass Support Vector Machines. IEEE Tran. on Neural Networks 13(2), 415–425 (2002)
Moro, R.A., Auria, L.: Support Vector Machines (SVM) as a Technique for Solvency Analysis. German Institute for Economic Research (811) (2008)
Huang, J., Lu, J., Ling, C.X.: Comparing naive Bayes, decision trees, and SVM with AUC and accuracy. In: Third IEEE International Conference on Data Mining, ICDM 2003, pp. 553–556 (2003)
Dimitoglou, G., Adams, J.A., Jim, C.M.: Comparison of the C4.5 and a Naive Bayes Classifier for the Prediction of Lung Cancer Survivability. Journal of Computing 4(8) (2012), http://www.journalofcomputing.org/valume-4-issue-8-august-2012
Anthony, Gregg, Tshilidzi: Image Classification Using SVMs: One-Against-One vs One-Against-All. In: 28th Asian Conference on Remote Sensing (2007)
Peng, C.J., Lee, K.L., Ingersoll, G.M.: An Introduction to Logistic Regression and Analysis Reporting. Journal of Educational Research (2002)
Kotsiantis, S.B.: Supervised Machine Learning: A Review of Classification Techniques. Informatica 31, 249–268 (2007)
Fletcher, T.: Support Vector Machines Explained (2009), http://www.tristanfletcher.co.uk
Hall, M., Frank, E., Holmes, G., Pfahringer, B., Reutemann, P., Witten, I.H.: The WEKA Data Mining Software: An Update. SIGKDD Explorations 11 (2009)
Fayyad, U., Irani, K.: Multi-interval Discretization of Continuous-valued attributes for Classification Learning. In: Proceeding of Thirteenth International Joint Conference on Artificial Intelligence, pp. 1022–1027. Morgan Kaufmann (1993)
Beygelzimer, A., Langford, J., Zadrozny, B.: Weighted One-Against-All. In: Proceeding of the 20th National Conference on Artificial Intelligence, pp. 720–725. AAAI Press (2005)
Allwein, E., Schapire, R., Singer, Y.: Reducing Multiclass to Binary: A unifying approach for margin classifiers. Journal of Machine Learning Research, 113–141 (2000)
Quinlan, J.R.: C4.5: Programs for Machine Learning. Morgan Kaufmann, San Mateo (1993)
Witten, I.H., Frank, E., Hall, M.: Data Mining: Practical Machine Learning Tools and Techniques, 3rd edn. Morgan Kaufmann Publishing (2011)
John, G.H., Langley, P.: Estimating Continuous Distributions in Bayesian Classifiers. In: Eleventh Conference on Uncertainty in Artificial Intelligence, San Mateo, pp. 338–345 (1995)
Hilbe, J.M.: Logistic Regression Models. Chapman & Hall/CRC Press (2009)
Corinna, C., Vapnik, V.: Support-Vector Networks. Machine Learning (1995)
Tsoumakas, G., Katakis, I.: Multi Label Classification: An Overview. International Journal of Data Warehousing and Mining 3, 1–13 (2007)
Sulzmann, J.-N., Fürnkranz, J., Hüllermeier, E.: On Pairwise Naive Bayes Classifiers. In: Kok, J.N., Koronacki, J., Lopez de Mantaras, R., Matwin, S., Mladenič, D., Skowron, A. (eds.) ECML 2007. LNCS (LNAI), vol. 4701, pp. 371–381. Springer, Heidelberg (2007)
Zadrozny, B.: Reducing Multiclass to Binary by Coupling Probability Estimates. In: Advances in Neural Information Processing Systems (2002)
Hastie, T., Tibshirani, R.: Classification by pairwise coupling. The Annals of Statistics 26(2), 451–471 (1998)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2013 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Eichelberger, R.K., Sheng, V.S. (2013). An Empirical Study of Reducing Multiclass Classification Methodologies. In: Perner, P. (eds) Machine Learning and Data Mining in Pattern Recognition. MLDM 2013. Lecture Notes in Computer Science(), vol 7988. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-39712-7_39
Download citation
DOI: https://doi.org/10.1007/978-3-642-39712-7_39
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-39711-0
Online ISBN: 978-3-642-39712-7
eBook Packages: Computer ScienceComputer Science (R0)