Abstract
Recently, nonparallel support vector machine (NPSVM), a branch of support vector machines (SVMs), is developed and has attracted considerable interest. A kind of developed NPSVM, ν-nonparallel support vector machine (ν-NPSVM), which inherits the advantages of ν-support vector machine (ν-SVM) enables us to achieve higher classification accuracy and less time to tune for parameters. However, applications of ν-NPSVM to large data sets are seriously hampered by their excessive training time. In this paper, we use divide-and-conquer technique for large scale ν-nonparallel support vector machine (DC-νNPSVM) aiming at overcoming this burden. In the division step, we first divide the whole samples into several smaller subsamples and solve smaller subproblems using ν-NPSVM independently. We theoretically and experimentally show that objective function value, solutions, and support vectors solved by (DC-νNPSVM) are likely to those of the whole ν-NPSVM. In the conquer step, subsolutions from subproblems are used as an initial coordinate descent solver which converges to the optimal solution quickly. Moreover, multi-level (DC-νNPSVM) is adopted to balance the accuracy and efficiency. (DC-νNPSVM) can achieve higher accuracy by tuning the parameters in a smaller range and control the number of support vectors efficiently. Quantities of experiments show our (DC-νNPSVM) outperforms state-of-the-art SVM methods for large scale data sets.






Similar content being viewed by others
Explore related subjects
Discover the latest articles and news from researchers in related subjects, suggested using machine learning.References
Cortes C, Vapnik VN (1995) Support-vector networks. Mach Learn 20(3):273–297
Vapnik VN (1996) The nature of statistical learning theory. Springer, New York
Vapnik VN (1998) Statistical learning theory. Wiley, New York
Trafalis TB, Ince H (2000) Support vector machine for regression and applications to financial forecasting. In: Proceedings of the IEEE-INNS-ENNS international joint conference on neural networks, vol 6, pp 348–353
Schölkopf B, Smola AJ, Williamson RC, Bartlett PL (2000) New support vector algorithms. Neural Comput 12:1207–1245
Tian YJ, Zhang Q, Liu DL (2014) ν-Nonparallel support vector machine for pattern classification. Neural Comput Appl 25:1007–1020
Platt JC (1998) Fast training of support vector machines using sequential minimal optimization. In: Schölkopf B, Burges CJC, Smola AJ (eds) Advances in Kernel Methods-Support Vector Learning. MIT Press, Cambridge
Chang CC, Lin CJ (2011) Libsvm: a linrary for support vector machines. ACM Trans Intell Syst Tecnol 2(27):1–27
Hsieh CJ, Chang KW, Lin CJ, Keerthi SS, Sundararajan S (2008) A dual coordinate descent method for large-scale linear svm. In: Proceedings of the 25th international conference on machine learning, pp 408–415
Fine S, Scheinberg K (2001) Efficient svm training using low-rank kernel representations. J Mach Learn Res 2:243–264
Zhang K, Lan L, Wang Z, Moerchen F (2012) Scaling up kernel SVM on Limited resources: a low-rank linearization approach. In: Proceedings of 15th internation conference on artificial intelligence and statistics (AISTATS)
Le Q, Sarlós T, Smola A (2013) Fastfood-approximating kernel expansions in loglinear time. In: Proceedings of the 30th international conference on machine learning
John M, Christian JD (1989) Fast learning in networks of locally-tuned processing units. Neural Comput 1:281–294
Yu H, Yang J, Han J, Li X (2005) Making SVMs scalable to large data sets using hierarchical cluster indexing. Data Min Knowl Discov 11(3):295–321
Zhang K, Tsang IW, Kwok JT (2008) Improved Nyström low rank approximation and error analysis. In: Proceedings of the 25st international conference on machine learning
Jacobs RA, Jordan MI, Nowlan SJ, Hinton GE (1991) Adaptive mixtures of local experts. Neural Comput 3(1):79–87
Tresp V (2000) A bayesian committee machine. Neural Comput 12:2719–2741
Graf HP, Cosatto E, Bottou L, Durdanovic I, Vapnik VN (2005) Parallel support vector machines: the cascade SVM. In: Proceedings of conference on neural information processing systems (NIPS)
Hsieh CJ, Si S, Dhillon IS (2014) A divide-and-conquer solver for kernel support vector machines. In: Proceedings of the 31st international conference on machine learning
Tian YJ, Ju XC, Shi Y (2016) A divide-and-combine method for large scale nonparallel support vector machines. Neural Netw 75:12–21
Tsang IW, Kwok JT, Cheung PM (2005) Core vector machines: fast svm training on very large data sets. J Mach Learn Res 6:363–392
Tsang IW, Kocsor A, Kwok JT (2007) Simpler core vector machines with enclosing balls. In: Proceedings of the 24th international conference on machine learning
Wang D, Zhang B, Zhang P, Qiao H (2010) An online core vector machine with adaptive meb adjustment. Pattern Recognit 42(10):3468–3482
Rahimi A, Recht B (2007) Random features for large-scale kernel machines. In: Advances in neural information processing systems
Kelley J (1960) The cutting-plane method for solving convex programs. J Soc Ind Appl Math 8(4):703–712
Joachims T (2006) Training linear svms in linear time. In: Proceedings of the 12th ACM SIGKDD international conference on knowledge discovery and data mining
Franc V, Sonnenburg S (2008) Optimized cutting plane algorithm for support vector machines. In: Proceedings of the 25th international conference on machine learning, pp 320–327
Teo CH, Vishwanathan S, Smola A, Le QV (2010) Bundle methods for regularized risk minimization. J Mach Learn Res 11:311–365
Zhang T (2004) Solving large scale linear prediction problems using stochastic gradient descent algorithms. In: Proceedings of the 21st international conference on machine learning
Shalev-Shwartz S, Singer Y, Srebro N, Cotter A (2011) Pegasos: primal estimated sub-gradient solver for SVM. Math Program 127:3–30
Bordes A, Ertekin S, Weston J, Bottou L (2005) Fast kernel classifiers with online and active learning. J Mach Learn Res 6:1579–1619
Fan RE, Chen PH, Lin CJ (2005) Working set selection using second order information for training SVM. J Mach Learn Res 6:1889–1918
Radha C, Rong J, Timothy CH, Anil KJ (2011) Approximate kernel k-means: solution to large scale kernel clustering. In: ACM SIGKDD conference on knowledge discovery and data mining
Tian YJ, Qi ZQ, Ju XC, Shi Y, Liu XH (2013) Nonparallel support vector machines for pattern classification. IEEE Trans Cybern 44(7):1067–1079
Acknowledgments
This work has been partially supported by grants from National Natural Science Foundation of China (Nos.61472390, 11271361, 71331005), Major International (Regional) Joint Research Project (No.71110107026).
Author information
Authors and Affiliations
Corresponding author
Rights and permissions
About this article
Cite this article
Ju, X., Tian, Y. A divide-and-conquer method for large scale ν-nonparallel support vector machines. Neural Comput & Applic 29, 497–509 (2018). https://doi.org/10.1007/s00521-016-2574-3
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-016-2574-3