Abstract
TSVM is designed to solve binary classification problems with less computational overhead by finding two hyperplanes and has been widely used to solve real-world problems. However, in real scenarios, the data used for learning is always scattered in different institutions or users. At the same time, people pay more attention to the issue of personal privacy leakage. Due to the complex privacy protection issues, simply collecting all the data for model training is no longer acceptable. Federated learning has recently been proposed to solve this problem. It completes model training by sharing model parameter updates in the form of data that remains local. But there is still no algorithm for twin support vector machines under the framework of federated learning. Combining the characteristics of twin support vector machine and federated learning, this paper proposes a federated twin support vector machine algorithm (FTSVM) and extends the twin support vector machine based on stochastic gradient descent into a federated support vector machine. We propose a unique initialization algorithm and integration algorithm to ensure the accuracy of the algorithm and the effectiveness of privacy protection. Accuracy experiments are carried out on five datasets, and the accuracy is basically the same as that of the TSVM based on gradient descent. Ablation experiments show that as the number of participants increases, the accuracy of the FTSVM is significantly higher than the average accuracy of the Stand-alone TSVM. Time experiments show that the time overhead of FTSVM increases linearly with the number of participants. These experiments demonstrate the accuracy, effectiveness, and possibility of application in real-world scenarios of our proposed FTSVM.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Blake, C.: UCI repository of machine learning databases (1998). http://www.ics.uci.edu/~mlearn/MLRepository.html
Chen, X., Yang, J., Ye, Q., Liang, J.: Recursive projection twin support vector machine via within-class variance minimization. Pattern Recogn. 44(10–11), 2643–2655 (2011)
Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20(3), 273–297 (1995)
Hao, P.Y.: New support vector algorithms with parametric insensitive/margin model. Neural Netw. 23(1), 60–73 (2010)
Jajadeva, D., Khemchandani, R., Chandra, S.: Twin Support Vector Machines: Models, Extensions and Applications. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-46186-1
Khemchandani, R., Chandra, S., et al.: Twin support vector machines for pattern classification. IEEE Trans. Pattern Anal. Mach. Intell. 29(5), 905–910 (2007)
Konečnỳ, J., McMahan, H.B., Ramage, D., Richtárik, P.: Federated optimization: distributed machine learning for on-device intelligence. arXiv preprint arXiv:1610.02527 (2016)
Konečnỳ, J., McMahan, H.B., Yu, F.X., Richtárik, P., Suresh, A.T., Bacon, D.: Federated learning: strategies for improving communication efficiency. arXiv preprint arXiv:1610.05492 (2016)
Kumar, M.A., Gopal, M.: Least squares twin support vector machines for pattern classification. Expert Syst. Appl. 36(4), 7535–7543 (2009)
McMahan, B., Moore, E., Ramage, D., Hampson, S., Arcas, B.A.: Communication-efficient learning of deep networks from decentralized data. In: Artificial Intelligence and Statistics, pp. 1273–1282. PMLR (2017)
Peng, X.: TPMSVM: a novel twin parametric-margin support vector machine for pattern recognition. Pattern Recogn. 44(10–11), 2678–2692 (2011)
Shalev-Shwartz, S., Singer, Y., Srebro, N., Cotter, A.: Pegasos: primal estimated sub-gradient solver for SVM. Math. Program. 127(1), 3–30 (2011)
Wang, S., Chen, M., Saad, W., Yin, C.: Federated learning for energy-efficient task computing in wireless networks. In: ICC 2020–2020 IEEE International Conference on Communications (ICC), pp. 1–6. IEEE (2020)
Wang, Z., Shao, Y.H., Bai, L., Li, C.N., Liu, L.M., Deng, N.Y.: Insensitive stochastic gradient twin support vector machines for large scale problems. Inf. Sci. 462, 114–131 (2018)
Xu, Y., Yang, Z., Pan, X.: A novel twin support-vector machine with pinball loss. IEEE Trans. Neural Netw. Learn. Syst. 28(2), 359–370 (2016)
Yang, Q., Liu, Y., Chen, T., Tong, Y.: Federated machine learning: concept and applications. ACM Trans. Intell. Syst. Technol. (TIST) 10(2), 1–19 (2019)
Zeng, D., Liang, S., Hu, X., Xu, Z.: FedLAB: a flexible federated learning framework. arXiv preprint arXiv:2107.11621 (2021)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Switzerland AG
About this paper
Cite this paper
Yang, Z., Chen, X. (2022). Federated Twin Support Vector Machine. In: Yu, S., et al. Pattern Recognition and Computer Vision. PRCV 2022. Lecture Notes in Computer Science, vol 13534. Springer, Cham. https://doi.org/10.1007/978-3-031-18907-4_15
Download citation
DOI: https://doi.org/10.1007/978-3-031-18907-4_15
Published:
Publisher Name: Springer, Cham
Print ISBN: 978-3-031-18906-7
Online ISBN: 978-3-031-18907-4
eBook Packages: Computer ScienceComputer Science (R0)