Abstract
A new Smooth Support Vector Machine (SSVM) is proposed and is called NSSVM for short. Different from traditional SSVM that treats perturbation formulation of SVM, NSSVM treats standard 2-norm error soft margin SVM. Different from traditional SSVM that uses the 2-norm of the Lagrangian multipliers vector to roughly substitute that of the weight of the separating hyperplane, which makes the obtained smooth model unequal to the primal program; NSSVM takes into account the connotative relation between the primal and dual program to transform the original program to a new smooth one. Numerical experiments on several UCI datasets demonstrate that NSSVM has higher precisions than existing methods.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Vapnik, V.N.: The Nature of Statistical Learning Theory. Springer, New York (2000)
Burges, C.J.C.: A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery 2, 121–167 (1998)
Yu, H., Han, J., Chang, K.C.: PEBL: Positive-example based learning for Web page classification using SVM. In: Proc. 8th Int. Cong. Knowledge Discovery and Data Mining, Edmonton, Canada (2002)
Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines. Cambridge University Press, Cambridge (2000)
Chen, X., Qi, L., Sun, D.: Global and superlinear convergence of the smoothing Newton method and its application to general box constrained variational inequalities. Mathematics of Computation 67, 519–540 (1998)
Qi, L., Sun, D.: Smoothing Functions and Smoothing Newton Method for complementarity and Variational Ineauqlity Problems. Journal of Optimization Theory and Applications 113(1), 121–147 (2002)
Lee, Y.-J., Mangasarian, O.L.: SSVM: A Smooth Support Vector Machine for Classification. Computational Optimization and Applications 22(1), 5–21 (2001)
Hsu, C.-W., Lin, C.-J.: A simple decomposition method for support vector machines. Machine Learning 46, 291–314 (2002)
Yuan, Y.-b., Yan, J., Xu, C.-x.: Polynomial smooth support vector machine (PSSVM). Chinese Journal of Computers 28(1), 9–71 (2005) (in Chinese)
Lee, Y.J., Hsieh, W.F., Huang, C.M.: SSVR: A smooth support vector machine for insensitive regression. IEEE Transactions on Knowledge and Data Engineering 17(5), 5–22 (2005)
Yan-Feng, F., De-Xian, Z., Hua-Can, H.: Smooth SVM Research: A polynomial- based Approach. In: Proceedings of the Sixth Conference on Information, Communications & Signal Processing, pp. 1–5 (2007)
Jin-zhi, X., Jin-lian, H., Hua-qiang, Y., Tian-ming, H., Guang-ming, L.: Research on a New Class of Functions for Smoothing Support Vector Machines. Acta Electronica Sinica 35(2), 366–370 (2007)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2010 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Liang, J., Wu, D. (2010). A New Smooth Support Vector Machine. In: Wang, F.L., Deng, H., Gao, Y., Lei, J. (eds) Artificial Intelligence and Computational Intelligence. AICI 2010. Lecture Notes in Computer Science(), vol 6319. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-642-16530-6_32
Download citation
DOI: https://doi.org/10.1007/978-3-642-16530-6_32
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-642-16529-0
Online ISBN: 978-3-642-16530-6
eBook Packages: Computer ScienceComputer Science (R0)