Abstract
Taking motivation from v-insensitive twin support vector regression (v-TSVR) and the projection idea, this paper proposes a novel v-twin projection support vector regression model, called v-TPSVR. This v-TPSVR, based on v-TSVR, determines the regression function through a pair of nonparallel hyperplanes solved by two smaller sized quadratic programming problems (QPPs). The proposed v-TPSVR model also can automatically optimize the parameters \( \varepsilon_{1} \) and \( \varepsilon_{2} \) via the user specified parameters \( v_{1} \) and \( v_{2} \), which is the same as v-TSVR. But different from v-TSVR, v-TPSVR seeks a projection axis in each QPP such that the variance of the projected points is minimized, so the empirical correlation coefficient between each hyperplane and the projected inputs is maximized. Although the training speed of the proposed algorithm is similar to that of other compared algorithms, it is obvious that the introduction of the projection axis makes the number of SV less than that of v -TSVR under the same values of \( v_{1} \) and \( v_{2} \), it would lead to faster testing speed. In addition, The experimental results indicate that the proposed v-TPSVR obtains the better prediction performance than the popular ɛ-TSVR and v-TSVR.
Similar content being viewed by others
References
Balasundaram S, Meena Y (2016) Training primal twin support vector regression via unconstrained convex minimization. Appl Intell 44:931–955
Blake CL, Merz CJ (1998) UCI repository for machine learning databases, Department of Information and Computer Sciences, University of California, Irvine. http://www.ics.uci.edu/~mlearn/MLRepository.html. Accessed 6 Apr 2018
Brown MPS, Grundy WN, Lin D (2000) Knowledge-based analysis of microarray gene expression data by using support vector machine. Proc Natl Acad Sci USA 97(1):262–267
Burges C (1998) A tutorial on support vector machines for pattern recognition. Data Min Knowl Discov 2:121–167
Gupta D (2017) Training primal K-nearest neighbor based weighted twin support vector regression via unconstrained convex minimization. Appl Intell 47(3):962–991
Hao PY (2010) New support vector algorithms with parametric insensitive/margin model. Neural Netw 23:60–73
Jayadeva, Khemchandani R, Chandra S (2007) Twin support vector machines for pattern classification. IEEE Trans Pattern Anal Mach Intell 29(5):905–910
Joachims T, Ndellec C, Rouveriol C (1998) Text categorization with support vector machines: learning with many relevant features. In: European conference on machine learning no. 10, Chemnitz, Germany pp 137–142
Khemchandani R, Saigal P, Chandra S (2016) Improvements on nu-twin support vector machine. Neural Netw 79:97–107
Lee YJ, Hsieh WF, Huang CM (2005) ɛ-SSVR: a smooth support vector machine for ɛ-insensitive regression. IEEE Trans Knowl Data Eng 17(5):678–685
Mehrkanoon S, Huang XL, Suykens JAK (2014) Non-parallel support vector classifiers with different loss functions. Neuro Comput 143:294–301
Osuna E, Freund R, Girosi F (1997) Training support vector machines: an application to face detection. In: Proceedings of IEEE computer vision and pattern recognition, San Juan, Puerto Rico, pp 130–136
Peng XJ (2010) TSVR: an efficient twin support vector machine for regression. Neural Netw 23:365–372
Peng XJ (2012) Efficient twin parametric insensitive support vector regression model. Neuro Comput 79:26–38
Peng XJ, Chen D (2018) PTSVRs: regression models via projection twin support vector machine. Inf Sci 435:1–14
Peng XJ, Xu D, Shen JD (2014) A twin projection support vector machine for data regression. Neuro Comput 138:131–141
Qi ZQ, Tian YJ, Shi Y (2013) Robust twin support vector machine for pattern classification. Pattern Recognit 46(1):305–316
Rastogi R, Anand P, Chandra S (2017) A v-twin support vector machine based regression with automatic accuracy control. Appl Intell 46:670–683
Scholkopf B, Smola AJ, Williamson RC, Bartlett PL (2000) New support vector algorithms. Neural Comput 12(5):1207–1245
Shao YH, Zhang CH, Wang XB, Deng NY (2011) Improvements on twin support vector machines. IEEE Trans Neural Netw 22(6):962–968
Shao YH, Zhang CH, Yang ZM, Jing L, Deng NY (2013) An ε-twin support vector machine for regression. Neural Comput Appl 23:175–185
Shao YH, Deng NY, Yang ZM (2014) Least squares recursive projection twin support vector machine for classification. Pattern Recogn 45(6):2299–2307
Shevade SK, Keerthi SS, Bhattacharyya C (2000) Improvements to the SMO algorithm for SVM regression. IEEE Trans Neural Netw 11(5):1188–1193
Tang L, Tian YJ, Yang CY (2018) Nonparallel support vector regression model and its SMO-type solver. Neural Netw 105:431–446
Tanveer M, Shubham K, Aldhaifallah M (2016a) An efficient regularized K-nearest neighbor based weighted twin support vector regression. Knowl-Based Syst 94:70–87
Tanveer M, Shubham K, Aldhaifallah M, Nisar KS (2016b) An efficient implicit regularized Lagrangian twin support vector regression. Appl Intell 44(4):831–848
Vapnik V (1998) Statistical learning theory. Wiley, New York
Wang Z, Shao YH, Bai L, Li CN, Liu LM, Deng NY (2018) Insensitive stochastic gradient twin support vector machines for large scale problems. Inf Sci 462:114–131
Acknowledgements
This work was partially supported by the National Natural Science Foundation of China (Nos. 61304002 and 61403177). In this paper, we also utilized a few public databases. We therefore thank the providers of these databases.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interests
We declare that there is no conflict of interest regarding the publication of this paper.
Additional information
Publisher’s Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations
Rights and permissions
About this article
Cite this article
Zhao, Nn., Ouyang, Xy., Gao, C. et al. A v-twin projection SVR with automatic accuracy adjustment. Artif Intell Rev 53, 1511–1527 (2020). https://doi.org/10.1007/s10462-019-09711-w
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10462-019-09711-w