Abstract
The reduced support vector machine was proposed for the practical objective that overcomes the computational difficulties as well as reduces the model complexity by generating a nonlinear separating surface for a massive dataset. It has been successfully applied to other kernel-based learning algorithms. Also, there are experimental studies on RSVM that showed the efficiency of RSVM. In this paper we propose a robust method to build the model of RSVM via RBF (Gaussian) kernel construction. Applying clustering algorithm to each class, we can generate cluster centroids of each class and use them to form the reduced set which is used in RSVM. We also estimate the approximate density for each cluster to get the parameter used in Gaussian kernel. Under the compatible classification performance on the test set, our method selects a smaller reduced set than the one via random selection scheme. Moreover, it determines the kernel parameter automatically and individually for each point in the reduced set while the RSVM used a common kernel parameter which is determined by a tuning procedure.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
Preview
Unable to display preview. Download preview PDF.
References
Burges, C.J.C.: A tutorial on support vector machines for pattern recognition. Data Mining and Knowledge Discovery 2(2), 121–167 (1998)
Chapelle, O., Vapnik, V., Bousquet, O., Mukherjee, S.: Choosing multiple parameters for support vector machines. Machine Learning 46(1), 131–159 (2002)
Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines. Cambridge University Press, Cambridge (2000)
Lee, Y.-J., Mangasarian, O.L.: RSVM: Reduced support vector machines. Technical Report 00-07, Data Mining Institute, Computer Sciences Department, University of Wisconsin, Madison, Wisconsin (July 2000); Proceedings of the First SIAM International Conference on Data Mining, Chicago, April 5-7 (2001), CDROM Proceedings, ftp://ftp.cs.wisc.edu/pub/dmi/tech-reports/00-07.ps
Lee, Y.-J., Mangasarian, O.L.: SSVM: A smooth support vector machine. Computational Optimization and Applications 20, 5–22 (2001); Data Mining Institute, University of Wisconsin, Technical Report 99-03, ftp://ftp.cs.wisc.edu/pub/dmi/tech-reports/99-03.ps
Lyhyaoui, A., Martinez, M., Mora, I., Vazquez, M., Sancho, J.-L., Figueiras-Vidal, A.R.: Sample selection via clustering to construct support vector-like classifier. IEEE Transactions on Neural Networks 10, 1474–1481 (1999)
Mangasarian, O.L.: Generalized support vector machines. In: Smola, A., Bartlett, P., Schölkopf, B., Schuurmans, D. (eds.) Advances in Large Margin Classifiers, pp. 135–146. MIT Press, Cambridge (2000), ftp://ftp.cs.wisc.edu/mathprog/tech-reports/98-14.ps
Murphy, P.M., Aha, D.W.: UCI machine learning repository (1992), www.ics.uci.edu/~mlearn/MLRepository.html
Oyang, Y.-J., Hwang, S.-C., Ou, Y.-Y., Chen, C.-Y., Chen, Z.-W.: An novel learning algorithm for data classification with radial basis function networks. In: Proceeding of 9th International Conference on Neural Information Processing, Singapore, November 2001, pp. 18–22 (2001)
Vapnik, V.N.: The Nature of Statistical Learning Theory. Springer, New York (1995)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2004 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Jen, LR., Lee, YJ. (2004). Clustering Model Selection for Reduced Support Vector Machines. In: Yang, Z.R., Yin, H., Everson, R.M. (eds) Intelligent Data Engineering and Automated Learning – IDEAL 2004. IDEAL 2004. Lecture Notes in Computer Science, vol 3177. Springer, Berlin, Heidelberg. https://doi.org/10.1007/978-3-540-28651-6_106
Download citation
DOI: https://doi.org/10.1007/978-3-540-28651-6_106
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-22881-3
Online ISBN: 978-3-540-28651-6
eBook Packages: Springer Book Archive