Abstract
SVM has been receiving increasing interest in areas ranging from its original application in pattern recognition to other applications such as regression estimation due to its remarkable generalization performance. However, it also contains some defects such as storage problem (in training process) and sparsity problem. In this paper, a new method is proposed to pre-select the base vectors from the original data according to vector correlation principle, which could greatly reduce the scale of the optimization problem and improve the sparsity of the solution. The method could capture the structure of the data space by approximating a basis of the subspace of the data; therefore, the statistical information of the training samples is preserved. In the paper, the process of mathematical deduction is given in details and results of simulations on artificial data and practical data have been done to validate the performance of base vector selection (BVS) algorithm. The experimental results show the combination of such algorithm with SVM can make great progress while can’t sacrifice the SVM’s performance.
This is a preview of subscription content, log in via an institution.
Buying options
Tax calculation will be finalised at checkout
Purchases are for personal use only
Learn about institutional subscriptionsPreview
Unable to display preview. Download preview PDF.
References
Cortes, C., Vapnik, V.: Support vector network. Mach. Learn. 20, 273–297 (1995)
Vapnik, V.: An overview of statistical learning theory. IEEE Trans. Neural Network 10(5), 988–999 (1999)
Burges, C.J.C.: A tutorial on support vector machines for pattern recognition. Data Mining Knowledge Discovery 2(2), 121–167 (1998)
Scholkopf, B., Smola, A.: Learning with Kernels. MIT Press, Cambridge (1999)
Platt, J.C.: Fast training of support vector machines using sequential minimal optimization. In: Smola, A.J., Burges, C.J.C., Schölkopf, B. (eds.) Advances in Kernel Methods-Support Vector Learning, MIT Press, Cambridge (1998)
Vapnik, V.: Three remarks on support vector machine. In: Solla, S.A., Leen, T.K., Müller, K.R. (eds.) Advances in Neural Comput., vol. 10, pp. 1299–1319 (1998)
Baudat, G., Anouar, F.: Generalized Discriminant Analysis Using a Kernel Approach. Neural Computation 12(10), 2385–2404 (2000)
Lang, K.J., Witbrock, M.J.: Learning to tell two spirals apart. In: Proc. 1989 Connectionist Models Summer School, pp. 52–61 (1989)
Author information
Authors and Affiliations
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2006 Springer-Verlag Berlin Heidelberg
About this paper
Cite this paper
Li, Q. (2006). Base Vector Selection for Support Vector Machine. In: Jiao, L., Wang, L., Gao, Xb., Liu, J., Wu, F. (eds) Advances in Natural Computation. ICNC 2006. Lecture Notes in Computer Science, vol 4221. Springer, Berlin, Heidelberg. https://doi.org/10.1007/11881070_28
Download citation
DOI: https://doi.org/10.1007/11881070_28
Publisher Name: Springer, Berlin, Heidelberg
Print ISBN: 978-3-540-45901-9
Online ISBN: 978-3-540-45902-6
eBook Packages: Computer ScienceComputer Science (R0)