Abstract
The classification speed of SVM is inversely proportional to the number of Support Vectors (SVs). Therefore, the less SVs means the more sparseness and the higher classification speed. In order to reduce the number of SVs but without losing of generalization performance, a new algorithm called Classification Algorithm of Support Vector Machine based on Adaptive Affinity Propagation clustering Granulation (CSVM-AAPG) is proposed, which employs Affinity Propagation (AP) clustering algorithm to cluster the original SVs and the cluster centers are used as the new SVs, then aiming to minimize the classification gap between SVM and CSVM-AAPG, a quadratic programming model is built for obtaining the optimal coefficients of the new SVs. Meanwhile, it is proven that when clustering the original SVs, the minimal upper bound of the error between the original decision function and the fast decision function can be achieved by AP. Finally, experiments show that compared with original SVs, the number of SVs decreases and the speed of classification increases using CSVM-AAPG, while the loss of accuracy is in the acceptable level.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Bellocchio, F., Ferrari, S., Piuri, V.: Hierarchical Approach for Multiscale Support Vector Regression. IEEE Transactions on Neural Networks and Learning Systems 23(9), 1448–1460 (2012)
Kumar, M.M., Khemchandani, R., Gopal, M., Chandra, S.: Knowledge Based Least Squares Twin Support Vector Machines. Information Sciences 180(23), 4606–4618 (2010)
Ye, Q.L., Zhao, C.X., Chen, X.B.: A Feature Selection Method for TWSVM via a Regularization Technique. Journal of Computer Research and Development 48(6), 1029–1037 (2011)
Shao, Y.H., Zhang, C.H., Wang, X.B., et al.: Improvements on Twin Support Vector Machines. IEEE Transactions on Neural Networks 22(6), 962–968 (2011)
Platt, J.C.: Fast Training of Support Vector Machines Using Sequential Minimal Optimization. In: Advances in Kernel Methods-Support Vector Learning, California, USA, pp. 185–208 (1999)
Ye, Q.L., Zhao, C.X., Ye, N., et al.: Localized Twin SVM via Convex Minimization. Neurocomputing 74, 580–587 (2011)
Xu, Y.T., Wang, L.S., Zhong, P.: A Rough Margin-based V-twin Support Vector Machine. Neural Computing & Applications 21, 1307–1317 (2012)
Qi, Z.Q., Tian, Y.J., Shi, Y.: Twin Support Vector Machine with Universum Data. Neural Networks 36, 112–119 (2012)
Zhang, X.S., Gao, X.B.: Twin Support Vector Machines and Subspace Learning Methods for Micro-calcification Clusters Detection. Engineering Applications of Artificial Intelligence 25, 1062–1072 (2012)
Shao, Y.H., Deng, N.Y.: A Novel Margin-based Twin Support Vector Machine with Unitynorm Hyperplanes. Neural Computing & Applications 22, 1627–1635 (2013)
Zhang, Z.C., Wang, S.T., et al.: Fast Decision Using SVM for Incoming Samples. Journal of Electronics & Information Technology 33(9), 2181–2186 (2011)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2014 Springer International Publishing Switzerland
About this paper
Cite this paper
Wei, X. (2014). A Reduction SVM Classification Algorithm Based on Adaptive AP Clustering Granulation. In: Huang, DS., Jo, KH., Wang, L. (eds) Intelligent Computing Methodologies. ICIC 2014. Lecture Notes in Computer Science(), vol 8589. Springer, Cham. https://doi.org/10.1007/978-3-319-09339-0_15
Download citation
DOI: https://doi.org/10.1007/978-3-319-09339-0_15
Publisher Name: Springer, Cham
Print ISBN: 978-3-319-09338-3
Online ISBN: 978-3-319-09339-0
eBook Packages: Computer ScienceComputer Science (R0)