Abstract
A framework of Genetic Algorithm-Support Vector Machine (GA-SVM) is proposed for SVM parameters (model) selection, and clustering algorithm is also integrated with the framework to generate multiple optimal models, as well as being condition of convergence for GA. Moreover, an ensemble method on various SVM models assisted by TRUST-TECH methodology is put forward, to enhance the generalization ability of a single SVM model. The performance of GA-SVM and ensemble method is testified by applying them in both classification and regression problems. Results show that, comparing with traditional parameters selection method (such as grid search), the proposed GA-SVM framework and ensemble strategy can solve general classification and regression issues more efficiently and automatically with better performance.
Access this chapter
Tax calculation will be finalised at checkout
Purchases are for personal use only
References
Cortes, C., Vapnik, V.: Support-vector networks. Mach. Learn. 20(3), 273–297 (1995)
Vapnik, V.: The Nature of Statistical Learning Theory. Springer, New York (2000). https://doi.org/10.1007/978-1-4757-3264-1
Vapnik, V.: An overview of statistical learning theory. IEEE Trans. Neural Netw. 10(5), 988–999 (1999)
Platt, J.: Sequential minimal optimization: a fast algorithm for training support vector machines (1998)
Vapnik, V., Chapelle, O.: Bounds on error expectation for support vector machines. Neural Comput. 12(9), 2013–2036 (2000)
Chang, M.W., Lin, C.J.: Leave-one-out bounds for support vector regression model selection. Neural Comput. 17(5), 1188–1222 (2005)
Chapelle, O., Vapnik, V.: Model selection for support vector machines. In: Advances in Neural Information Processing Systems, pp. 230–236 (1999)
Hsu, C.W., Chang, C.C., Lin, C.J.: A practical guide to support vector classification (2003)
Huang, C.L., Wang, C.J.: A GA-based feature selection and parameters optimization for support vector machines. Expert Syst. Appl. 31(2), 231–240 (2006)
Pai, P.F., Hong, W.C.: Support vector machines with simulated annealing algorithms in electricity load forecasting. Energy Convers. Manag. 46(17), 2669–2688 (2005)
Chapelle, O., Vapnik, V., Bousquet, O., Mukherjee, S.: Choosing multiple parameters for support vector machines. Mach. Learn. 46(1), 131–159 (2002)
Dietterich, T.G.: Ensemble methods in machine learning. In: Kittler, J., Roli, F. (eds.) MCS 2000. LNCS, vol. 1857, pp. 1–15. Springer, Heidelberg (2000). https://doi.org/10.1007/3-540-45014-9_1
Kim, H.C., Pang, S., Je, H.M., Kim, D., Bang, S.Y.: Constructing support vector machine ensemble. Pattern Recogn. 36(12), 2757–2767 (2003)
Zhang, Y.F., Chiang, H.D.: A novel consensus-based particle swarm optimization-assisted trust-tech methodology for large-scale global optimization. IEEE Trans. Cybern. 47(9), 2717–2729 (2017)
Zhang, Y.F., Chiang, H.D., Wang, T.: A novel TRUST-TECH-enabled trajectory-unified methodology for computing multiple optimal solutions of constrained nonlinear optimization: theory and computation. IEEE Trans. Syst. Man Cybern. Syst. 52(1), 473–484 (2022)
Nguyen, T., Gordon-Brown, L., Wheeler, P., Peterson, J.: GA-SVM based framework for time series forecasting. In: 5th International Conference on Natural Computation, pp. 493–498. IEEE (2009)
Wang, L., Xu, G., Wang, J., Yang, S., Guo, L., Yan, W.: GA-SVM based feature selection and parameters optimization for BCI research. In: 7th International Conference on Natural Computation, pp. 580–583. IEEE (2011)
Cristianini, N., Shawe-Taylor, J.: An Introduction to Support Vector Machines and Other Kernel-Based Learning Methods. Cambridge University Press, Cambridge (2000)
Chang, C.C., Lin, C.J.: LIBSVM: a library for support vector machines. ACM Trans. Intell. Syst. Technol. (TIST) 2(3), 1–27 (2011)
Srinivas, M., Patnaik, L.M.: Genetic algorithms: a survey. Computer 27(6), 17–26 (1994)
Ball, G. H., Hall, D. J.: ISODATA, a novel method of data analysis and pattern classification. Stanford Research Inst., Menlo Park, CA (1965)
Wang, B., Chiang, H.D.: ELITE: ensemble of optimal input-pruned neural networks using TRUST-TECH. IEEE Trans. Neural Netw. 22(1), 96–109 (2007)
Zhang, Y.F., Chiang, H.D.: Enhanced ELITE-load: a novel CMPSOATT methodology constructing short-term load forecasting model for industrial applications. IEEE Trans. Ind. Inf. 16(4), 2325–2334 (2020)
Reddy, C.K., Chiang, H.D., Rajaratnam, B.: Trust-tech-based expectation maximization for learning finite mixture models. IEEE Trans. Pattern Anal. Mach. Intell. 30(7), 1146–1157 (2008)
Chiang, H.D., Wang, B., Jiang, Q.Y.: Applications of TRUST-TECH methodology in optimal power flow of power systems. In: Kallrath, J., Pardalos, P.M., Rebennack, S., Scheidt, M. (eds.) Optimization in the Energy Industry, pp. 297–318. Springer, Heidelberg (2009). https://doi.org/10.1007/978-3-540-88965-6_13
Rao, N.S.V.: On fusers that perform better than best sensor. IEEE Trans. Pattern Anal. Mach. Intell. 23(8), 904–909 (2001)
Zhou, Z.H., Wu, J., Tang, W.: Ensembling neural networks: many could be better than all. Artif. Intell. 137(1–2), 239–263 (2002)
Opitz, D.W., Shavlik, J.W.: Generating accurate and diverse members of a neural-network ensemble. In: Advances in Neural Information Processing Systems, pp. 535–541 (1995)
Brown, G.: Diversity in neural network ensembles. University of Birmingham (2004)
Windeatt, T.: Accuracy/diversity and ensemble MLP classifier design. IEEE Trans. Neural Netw. 17(5), 1194–1211 (2006)
Hashem, S.: Optimal linear combinations of neural networks. Neural Netw. 10(4), 599–614 (1997)
Ueda, N.: Optimal linear combination of neural networks for improving classification performance. IEEE Trans. Pattern Anal. Mach. Intell. 22(2), 207–215 (2000)
Sapankevych, N.I., Sankar, R.: Time series prediction using support vector machines: a survey. IEEE Comput. Intell. Mag. 4(2), 24–38 (2009)
Chen, B.J., Chang, M.W.: Load forecasting using support vector machines: a study on EUNITE competition 2001. IEEE Trans. Power Syst. 19(4), 1821–1830 (2004)
Author information
Authors and Affiliations
Corresponding author
Editor information
Editors and Affiliations
Rights and permissions
Copyright information
© 2022 The Author(s), under exclusive license to Springer Nature Singapore Pte Ltd.
About this paper
Cite this paper
Zhang, YF., Chiang, HD., Qu, YF., Zhang, X. (2022). TRUST-TECH Assisted GA-SVM Ensembles and Its Applications. In: Zhang, H., et al. Neural Computing for Advanced Applications. NCAA 2022. Communications in Computer and Information Science, vol 1637. Springer, Singapore. https://doi.org/10.1007/978-981-19-6142-7_8
Download citation
DOI: https://doi.org/10.1007/978-981-19-6142-7_8
Published:
Publisher Name: Springer, Singapore
Print ISBN: 978-981-19-6141-0
Online ISBN: 978-981-19-6142-7
eBook Packages: Computer ScienceComputer Science (R0)