Abstract
Multiple birth support vector machine (MBSVM) is a new classification algorithm, which includes the advantages of low complexity and high computing efficiency. However, the traditional MBSVM does not take into account the correlation sequence information among all dimensions of the samples when using the method to classify datasets, which limits the further improvement of the classification accuracy. Although some scholars have combined neural networks with support vector machine (SVM), these methods do not take into account the sequence correlation among different features. For the above problems, we present several variants of MBSVM algorithms to illustrate the validity and reliability of the theory: Multiple Birth Support Vector Machine based on Multilayer Perceptron (MLP-MBSVM), Multiple Birth Support Vector Machine based on Long-Short Term Memory Networks (LSTM-MBSVM), Multiple Birth Support Vector Machine based on Multilayer Perceptron and Long-Short Term Memory Networks(MLP-LSTM-MBSVM). After introducing multilayer perceptron and long-short term memory networks, these algorithms can take full account of the sequence correlation information between different features of samples. The experiments results show that the algorithms proposed in this paper are effective, and they can greatly improve the classification accuracy of multiple birth support vector machine.
Similar content being viewed by others
References
Vapnik VN (1997) The nature of statistical learning theory. IEEE Trans Neural Netw 38(4):409–409
Vapnik VN (1997) The support vector method. In: Proceedings of the 7th International conference on artificial neural networks, Lausanne, pp 263–271
Cristianini N, Shawe-Taylor J (2001) An introduction to support vector machines and other kernel-based learning methods. Kybernets 30(1):103–115
Ding SF, Yu JZ, Qi BJ et al (2014) An overview on twin support vector machines. Artif Intell Rev 42(2):245–252
An YX, Ding SF, Shi SH et al (2018) Discrete space reinforcement learning algorithm based on support vector machine classification. Pattern Recogn Lett 111:30–35
Ding SF, Shi ZZ, Tao DC et al (2016) Recent advances in support vector machines. Neurocomputing 211:1–3
Jayadeva, Reshma K, Chandra S (2007) Twin support vector machines for pattern classification. IEEE Trans Pattern Anal Mach Intell 29(5):905–910
Hua XP, Ding SF (2015) Weighted least squares projection twin support vector machines with local information. Neurocomputing 160:228–237
Ding SF, Huang HJ, Xu XZ et al (2014) Polynomial smooth twin support vector machines. Applied Mathematics & Information Sciences 8(4):2063–2071
Ding SF, Hua XP (2014) Recursive least squares projection twin support vector machines. Neurocomputing 130:3–9
Reshma R, Sharma S (2019) Fast laplacian twin support vector machine with active learning for pattern classification. Appl Soft Comput 74:424–439
Ding SF, An YX, Zhang XK et al (2016) Wavelet twin support vector machines based on glowworm swarm optimization. Neurocomputing 225(C):157–163
Zhang XK, Ding SF, Xue Y (2017) An improved multiple birth support vector machine for pattern classification. Neurocomputing 225:119–128
Yang ZX, Shao YH, Zhang XS (2013) Multiple birth support vector machine for multi-class classification. Neural Comput & Applic 22(1):153–161
Ding SF, Zhao XY, Zhang J et al (2019) A review on multi-class TWSVM. Artif Intell Rev 52(2):775–801
Ding SF, Zhang XK, An YX et al (2017) Weighted linear loss multiple birth support vector machine based on information granulation for multi-classclassification. Pattern Recogn 67:32–46
Ding SF, Huang HJ, Yu ZJ et al (2015) Research on the hybrid models of granular computing and support vector machine. Artif Intell Rev 43(4):565–577
Ding SF, Zhang XK (2016) Granular multiple birth support vector machine based on weighted linear loss. In: Proceedings of 2016 International Joint Conference on Neural Networks, Vancouver, pp 2453–2459
Chen S, Wu X (2017) Multiple birth least squares support vector machine for multi-class classification. Int J Mach Learn Cybern 8(6):1731–1742
Rosenblatt F (1958) The perception: a probabilistic model for information storage and organization in the brain. Psychol Rev 65(6):386–408
Minsky ML, Papert S (1970) A review of "Perceptrons: An introduction to computational geometry". Inf Control 17(5):501–522
Rumelhart DE, Williams D, Hinton G et al (1986) Learning representations by back-propagating errors. Nature 323(6088):533–536
Hinton GE, Salakhutdinov RR (2006) Reducing the dimensionality of data with neural networks. Science 313(5786):504–507
Glorot X, Bordes A, Bengio Y et al (2011) Deep Sparse Rectifier Neural Networks. In: Proceedings of the 14th International Conference on Artificial Intelligence and Statistics, Fort Lauderdale, pp 315–323
Lecun Y, Bengio Y, Hinton G (2015) Deep learning. Nature 521(7553):436–444
Silver D, Schrittwieser J, Simonyan K et al (2017) Mastering the game of go without human knowledge. Nature 550(7676):354–359
Sui XD, Zheng YJ, Wei BZ et al (2017) Choroid segmentation from optical coherence tomography with graph-edge weights learned from deep convolutional neural networks. Neurocomputing 237:332–341
Wang QC, Zheng YJ, Yang GP et al (2017) Multi-scale rotation-invariant convolutional neural networks for lung texture classification. IEEE Journal of Biomedical and Health Informatics:1–1
Zeng K, Ding SF (2019) Single image super-resolution using a polymorphic parallel CNN. Appl Intell 49(1):292–300
Bian WX, Ding SF, Zhang N et al (2019) Combined filtering and DBM reconstructing for fingerprint enhancement. Journal of Software 30(6):1886–1900
Hopfield J (1982) Neural networks and physical systems with emergent collective computational abilities. Proc Natl Acad Sci USA 79(8):2554–2558
Zhang J, Ding SF, Zhang N et al (2016) An incremental extreme learning machine based on deep feature embedded. International Journal of Machine Learning andCybernetics 7(1):111–120
Ding SF, Zhang N, Xu XZ et al (2015) Deep extreme learning machine and its application in EEG classification. Mathematical Problems in Engineering 2015:129021 11 pages
Ding SF, Guo LL, Hou YL (2017) Extreme learning machine with kernel model based on deep learning. Neural Comput & Applic 28(8):1975–1984
Zhang N, Ding SF, Zhang J et al (2017) Research on point wise gated deep networks. Appl Soft Comput 52:1210–1221
Hochreiter S, Schmidhuber J (1997) Long short-term memory. Neural Comput 9(8):1735–1780
Zhang Y, Chen G, Yu D et al (2016) Highway long short-term memory RNNS for distant speech recognition. IEEE:5755–5759
Wand M, Koutník J, Schmidhuber J (2016) Lipreading with long short-term memory. IEEE:6115–6119
Graves A (2012) Long short-term memory. Supervised sequence labelling with recurrent neural networks. Springer Berlin Heidelberg:1735–1780
Wiering MA, Schutten M, Millea A et al (2013) Deep support vector machines for regression problems. The Workshop on Advances in Regularization:141–148
Qi Z, Wang B, Tian Y et al (2016) When ensemble learning meets deep learning: a new deep support vector machine for classification. Knowl-Based Syst 107:54–60
Alalshekmubarak A, Smith L. S (2013) A novel approach combining recurrent neural network and support vector machines for time series classification. In: Proceedings of the 9th International Conference on Innovations in Information Technology (IIT), Abu Dhabi, pp 42–47
Melki G, Kecman V, Ventura S et al (2018) OLLAWV: online learning algorithm using worst-violators. Appl Soft Comput 66:384–393
Xu Y, Yang Z, Pan X et al (2017) A novel twin support vector machine with pinball loss. IEEE Transactions on Neural Networks and Learning Systems 28(2):359–370
Ju X, Tian Y, Liu D et al (2015) Nonparallel hyperplanes support vector machine for multi-class classification. Procedia Computer Science 51(1):1574–1582
Demišar, Janez, Schuurmans D (2006) Statistical Comparisons of classifiers over multiple data sets. J Mach Learn Res 7(1):1–30
Acknowledgements
This work is supported by the National Natural Science Foundation of China under Grant No.61672522, No.61976216, and No.61379101.
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declared that we have no conflicts of interest to this work.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Ding, S., Sun, Y., An, Y. et al. Multiple birth support vector machine based on recurrent neural networks. Appl Intell 50, 2280–2292 (2020). https://doi.org/10.1007/s10489-020-01655-x
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-020-01655-x