Abstract
One of the most challenging issues in pattern recognition is the data attribution selection process. Feature selection plays a key role in solving problems with high-dimensional data and is a fundamental step in pre-processing many classifications and machine learning problems. The feature selection method reduces the amount of data and increases the category precision. Unrelated data, which can lead to inappropriate classification, is thus removed to obtain fewer features. In this paper, the Binary Gray Wolf Optimization algorithm uses the Wrapper method for feature selection. The transfer function is an essential part of BGWO to map a continuous value to a binary value. In this study, eight transfer functions are divided into two families: S-shaped and V-shaped. Previous research has used only one transfer function for the whole algorithm, and all wolves in the whole algorithm deal with this transfer function. In this paper, each wolf has its own transfer function. Because algorithms are evolutionary meta-innovations and can optimize themselves, each wolf can play a role in the whole algorithm at any stage while optimizing itself and adapting to its community, and not just depend on one transfer function. In the proposed method, eight transfer functions are divided into two families, S-shaped and V-shaped. This article proposes two approaches for learning the transfer function, by selecting the transfer function and the slope of these functions. In the first approach, we add three or two binary bits to the initial population. If two bits are added, four modes of the transfer function are available, and if three bits are added, eight transfer functions are achievable. These bits are used as a criterion for selecting a predefined transfer function for each wolf. So, in the proposed method, each wolf has its transfer function. During the implementation of the algorithm, the wolves update their position according to the evaluation function and learning. In the second approach, ten or twenty-one binary bits are added to the initial population. If ten binary bits are used, we will have a transfer function, and \(2^{10}\) coefficient modes are available for the slope of the transfer function. If twenty-one binary bits are used, we have two transfer functions available. So, there are \(2^{10}\) modes for the gradient of the transfer function. These bits are used as a criterion for selecting the transfer function and the coefficient affecting the slope of these functions. In both ideas, after each iteration of the algorithm, the position of the wolves is updated and based on the evaluation function, the alpha wolf is identified and the transfer function is selected. With subsequent iterations, the algorithm learns and optimizes the transfer function to achieve the best feature selection with the smallest error. Experimental results on ten UCI datasets show that selecting the obtained feature subset with high classification accuracy is efficient.






Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Ziemba P, Piwowarski M (2011) Feature selection methods in data mining techniques, Prace Naukowe Uniwersytetu Ekonomicznego we Wrocławiu, 213–223
Chen R-C, Dewi C, Huang S-W, Caraka RE (2020) Selecting critical features for data classification based on machine learning methods. J Big Data 7:1–6
Zawbaa HM, Emary E, Grosan C, Snasel V (2018) Large-dimensionality small-instance set feature selection: a hybrid bio-inspired heuristic approach. Swarm Evolution Comput 42:29–42
Zamani H, Nadimi-Shahraki M-H (2016) Feature selection based on whale optimization algorithm for diseases diagnosis. Int J Comput Sci Inform Security 14(9):1243
Hsu H-H, Hsieh C-W, Lu M-D (2011) Hybrid feature selection by combining filters and wrappers. Expert Syst Appl 38:8144–8150
Gangavarapu T, Patil N (2019) A novel filter–wrapper hybrid greedy ensemble approach optimized using the genetic algorithm to reduce the dimensionality of high-dimensional biomedical datasets. Applied Soft Computing 81:105538
Hoque N, Bhattacharyya DK, Kalita JK (2014) MIFS-ND: a mutual information-based feature selection method. Expert Syst Appl 41(14):6371–6385
Xue B, Zhang M, Browne WN (2013) Particle swarm optimization for feature selection in classification: a multi-objective approach. IEEE Trans Cybern 43(6):1656–1671
Zhang R, Nie F, Li X, Wei X (2019) Feature selection with multiview data: a survey. Inform Fusion 50:158–167
Sánchez-Maroño N, Alonso-Betanzos A, Tombilla-Sanromán M (2007) Filter methods for feature selection–a comparative study. In: Intelligent data engineering and automated learning–ideal, pp 178–187
Nnamoko NA, Arshad FN, England D, Vora J, Norman J (2014) Evaluation of filter and wrapper methods for feature selection in supervised machine learning. In: Conference: the 15th annual postgraduate symposium on the convergence of telecommunication, networking and broadcasting at: liverpool
Cai J, Luo J, Wang S, Yang S (2018) Feature selection in machine learning: a new perspective. Neurocomputing 300:70–79
Venkatesh B, Anuradha J (2019) A review of feature selection and its methods. Cybern Inform Technol 19(1):3–26
Wu Y, Liu Y, Wang Y, Shi Y, Zhao X (2018) JCDSA: a joint covariate detection tool for survival analysis on tumor expression profiles. BMC Bioinform 19(1):1–8
Yang R, Zhang C, Zhang L, Gao R (2018) A two-step feature selection method to predict Cancerlectins by Multiview features and synthetic minority oversampling technique. BioMed Res Int. https://doi.org/10.1155/2018/9364182
Masoudi Sobhanzadeh Y, Motieghader H, Masoudi-Nejad A (2019) Feature Select: a software for feature selection based on machine learning approaches. BMC Bioinform 20(170):1–17
Metin SK (2018) Feature selection in multiword expression recognition. Expert Syst Appl 92(C):106–123
Saxena AK, Dubey VK, Wang J (2017) Hybrid feature selection methods for high-dimensional multi-class datasets. Int J Data Min Modell Manag 9(4):315
Bolon-Canedo V, Sanchez-Marono N, Alonso-Betanzos A (2012) An ensemble of filters and classifiers for microarray data classification. Pattern Recogn 45(1):531–539
Hoque N, Singh M, Bhattacharyya DK (2018) EFS-MI: an ensemble feature selection method for classification. Complex Intell Syst 4:105–118
Bolon-Canedo V, Sanchez-Marono N, Alonso-Betanzos A, Benitez JM, Herrera F (2014) A review of microarray datasets and applied feature selection methods. Inf Sci 282:111–135
Agrawal P, Abutarboush HF, Ganesh T, Mohamed AW (2021) Metaheuristic Algorithms on feature selection: A survey of one decade of research (2009–2019). IEEE Access 9:26766–26791
Jiang Y, Luo Q, Wei Y, Abualigah L, Zhou Y (2021) An efficient binary Gradient-based optimizer for feature selection. Math Biosci Eng 18(4):3813–3854
Emary E, Zawbaa HM, Hassanien AE (2016) Binary grey wolf optimization approaches for feature selection. Neuro Comput 172:371–381
Mirjalili S, Lewis A (2013) S-shaped versus V-shaped transfer functions for binary particle swarm optimization. Swarm Evol Comput 9:114
Mirjalili S, Yang X-S, Mirjalili SM (2014) Binary bat algorithm. Neurl Comput Appl 25(3):663–681
Mirjalili S, Mirjalili SM, Lewis A (2014) Grey wolf optimizer. Adv Eng Softw 69:46–61
Zhang J, Hong L, Liu Q (2021) An improved whale optimization algorithm for the traveling salesman problem. Symmetry 13(1):48
Hussien AG, Houssein EH, Hassanien AE (2017) A binary whale optimization algorithm with hyperbolic tangent fitness function for feature selection. In: Eighth international conference on intelligent computing and information systems (ICICIS)
Hu P, Pan JS, Chu SC (2020) Improved Binary Grey Wolf Optimizer and Its application for a feature selection. Knowledg Based Syst 195:105746
Altman NS (1992) An introduction to kernel and nearest-neighbor nonpara-metric regression. Am Statist 46(3):175–185
Ghosh M et al (2019) Genetic algorithm based cancerous gene identification from microarray data using ensemble of filter methods. Med Biol Eng Comput 57(1):159–176
Chellammal S, Sharmila R (2019) Recommendation of attributes for heart disease prediction using correlation measure. Int J Recent Technol Eng (IJRTE) 8(23):870–875
Author information
Authors and Affiliations
Contributions
Zahra Nassiri: Programmer Validation, Visualization, Investigation, Writing—original draft. Hesam Omranpour: Supervision, Project administration, Conceptualization, Methodology, Writing—reviewing and editing.
Corresponding author
Ethics declarations
Conflicts of interest
The authors declare that they have no known competing reported in this paper.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Nassiri, Z., Omranpour, H. Learning the transfer function in binary metaheuristic algorithm for feature selection in classification problems. Neural Comput & Applic 35, 1915–1929 (2023). https://doi.org/10.1007/s00521-022-07869-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-022-07869-z