Abstract
In this paper, we focus on multiple-class classification problems. By using polytomous logistic regression and support vector machine together, we come out a hybrid multi-class classifier with very promising results in terms of classification accuracy. Usually, the multiple-class classifier can be built by using many binary classifiers as its construction bases. Those binary classifiers might be trained by either one-versus-one or one-versus-others manners, and the final classifier is constructed by some kinds of “leveraging” methods; such as majority vote, weighted vote, regression, etc. Here, we propose a new way for constructing binary classifiers, which might take the relationship of classes into consideration. For example, the level of severity of a disease in medial diagnostic. Depending on the methods used for constructing binary classifiers, the final classifier will be constructed/assembled by nominal, ordinal or even more sophisticated polytomous logistic regression techniques. This hybrid method has been apply to many real world bench mark data sets and the results shows that this new hybrid method is very promising and out-performs the classifiers using the technique of the support vector machine alone.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Allwein, E., Schapire, R., Singer, Y.: Reducing Multiclass to Binary: A unifying Approach for Margin Classifiers. Journal of Machine Learning Research, 113–141 (2000)
Begg, C.B., Gray, R.: Calculation of polychotomous logistic regression parameters using individualized regression. Biometrika 71, 11–18 (1984)
Blake, C.L., Merz, C.J.: UCI repository of Machine Learning databases (1998)
Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines and Other Kernel-based Learning Methods. Cambridge University Press, Cambridge (2000)
Firth, D.: Bias reduction, the Jefferys prior and GLIM. In: Fahemeir, L., Francis, R.G., Tutz, G. (eds.) Advances in GLIM and Statistical Modelling, pp. 91–100. Springer, New York (1992)
Heinze, Schemper: A solution to the problem of separation in logistic regression. Statist. Med. 21, 2109–2149 (2002)
McCullagh, P.: Generalized Linear Models, 2nd edn. Chapman and Hall, New York (1989)
Rifkin, R., Klautau, A.: In defence of one-vs-all classification. Journal of Machine Learning Research, 101–141 (2004)
Schölkopf, B., Smola, A.J.: Learning with Kernels:Support Vector machines, Regularization, Optimization, and Beyond. MIT press, Cambridge (2001)
Vapnik, V.: The Nature of Statistical Learning Theory. Springer, NY (1995)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2004 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Chang, Yc.I., Lin, SC. (2004). Synergy of Logistic Regression and Support Vector Machine in Multiple-Class Classification. In: Yang, Z.R., Yin, H., Everson, R.M. (eds) Intelligent Data Engineering and Automated Learning – IDEAL 2004. IDEAL 2004. Lecture Notes in Computer Science, vol 3177. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-28651-6_19
Download citation
DOI: https://doi.org/10.1007/978-3-540-28651-6_19
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-22881-3
Online ISBN: 978-3-540-28651-6
eBook Packages: Springer Book Archive