Abstract
Feature selection (FS) is used to solve hard optimization problems in artificial intelligence and data mining. In the FS process, some, rather than all of the features of a dataset are selected in order to both maximize classification accuracy and minimize the time required for computation. In this paper a FS wrapper method that uses K-nearest Neighbor (KSN) classification is subjected to two modifications using a current improvement algorithm, the Monarch Butterfly Optimization (MBO) algorithm. The first modification, named MBOICO, involves the utilization of an enhanced crossover operator to improve FS. The second, named MBOLF, integrates the Lévy flight distribution into the MBO to improve convergence speed. Experiments are carried out on 25 benchmark data sets using the original MBO, MBOICO and MBOLF. The results show that MBOICO is superior, so its performance is also compared against that of four metaheuristic algorithms (PSO, ALO, WOASAT, and GA). The results indicate that it has a high classification accuracy rate of 93% on average for all datasets and significantly reduces the selection size. Hence, the findings demonstrate that the MBOICO outperforms the other algorithms in terms of classification accuracy and number of features chosen (selection size).








Similar content being viewed by others
Explore related subjects
Discover the latest articles and news from researchers in related subjects, suggested using machine learning.References
Niu X, Wang J (2019) A combined model based on data preprocessing strategy and multi-objective optimization algorithm for short-term wind speed forecasting. Appl Energy 241:519–539
Masoudi-Sobhanzadeh Y, Motieghader H, Masoudi-Nejad A (2019) FeatureSelect: a software for feature selection based on machine learning approaches. BMC Bioinf 20(1):170
Sayed S, Nassef M, Badr A, Farag I (2019) A nested genetic algorithm for feature selection in high-dimensional cancer microarray datasets. Expert Syst Appl 121:233–243
Alweshah M, Qadoura MA, Hammouri AI, Azmi MS, AlKhalaileh S (2020) Flower Pollination Algorithm for Solving Classification Problems. Int. J. Advance Soft Compu. Appl 12(1)
Al Nsour H, Alweshah M, Hammouri AI, Al Ofeishat H, Mirjalili S (2018) A hybrid grey wolf optimiser algorithm for solving time series classification problems. J Intell Syst 29(1):846–857
Mohamed AW, Hadi AA, Jambi KM (2019) Novel mutation strategy for enhancing SHADE and LSHADE algorithms for global numerical optimization. Swarm and Evolutionary Computation 50:100455
Manoj RJ, Praveena MA, Vijayakumar K (2019) An ACO–ANN based feature selection algorithm for big data. Clust Comput 22(2):3953–3960
Alweshah M, Al-Sendah M, Dorgham OM, Al-Momani A, Tedmori S (2020) Improved water cycle algorithm with probabilistic neural network to solve classification problems. Clust Comput:1–16
M Alweshah, L Rababa, MH Ryalat, A Al Momani, and MF Ababneh (2020). “African Buffalo algorithm: training the probabilistic neural network to solve classification problems,” Journal of King Saud University-Computer and Information Sciences
M Alweshah, A Al-Daradkeh, MA Al-Betar, A Almomani, and S Oqeili (2019). “$$\beta $$ β-hill climbing algorithm with probabilistic neural network for classification problems,” J Ambient Intell Humaniz Comput, pp. 1–12
Hashemi A, Dowlatshahi MB, Nezamabadi-pour H (2020) MGFS: a multi-label graph-based feature selection algorithm via PageRank centrality. Expert Syst Appl 142:113024
H Djellali, NG Zine, and N Azizi (2016). “Two stages feature selection based on filter ranking methods and SVMRFE on medical applications,” In Modelling and Implementation of Complex Systems: Springer, pp. 281–293
Hancer E, Xue B, Zhang M (2018) Differential evolution for filter feature selection based on information theory and feature ranking. Knowl-Based Syst 140:103–119
Apolloni J, Leguizamón G, Alba E (2016) Two hybrid wrapper-filter feature selection algorithms applied to high-dimensional microarray experiments. Appl Soft Comput 38:922–932
Liu H, Sun J, Liu L, Zhang H (2009) Feature selection with dynamic mutual information. Pattern Recogn 42(7):1330–1339
Karegowda AG, Manjunath A, Jayaram M (2010) Comparative study of attribute selection using gain ratio and correlation based feature selection. International Journal of Information Technology and Knowledge Management 2(2):271–277
Uğuz H (2011) A two-stage feature selection method for text categorization by using information gain, principal component analysis and genetic algorithm. Knowl-Based Syst 24(7):1024–1032
J Tang, S Alelyani, and H Liu (2014). “Feature selection for classification: A review,” Data classification: Algorithms and applications, p. 37
Rodriguez-Galiano V, Luque-Espinar J, Chica-Olmo M, Mendes M (2018) Feature selection approaches for predictive modelling of groundwater nitrate pollution: an evaluation of filters, embedded and wrapper methods. Sci Total Environ 624:661–672
Yusta SC (2009) Different metaheuristic strategies to solve the feature selection problem. Pattern Recogn Lett 30(5):525–534
Wang J, Hedar A-R, Wang S, Ma J (2012) Rough set and scatter search metaheuristic based feature selection for credit scoring. Expert Syst Appl 39(6):6123–6128
Wahono RS, Suryana N, Ahmad S (2014) Metaheuristic optimization based feature selection for software defect prediction. Journal of Software 9(5):1324–1333
Alweshah M, Rashaideh H, Hammouri AI, Tayyeb H, Ababneh M (2017) Solving time series classification problems using support vector machine and neural network. International journal of data analysis techniques and strategies 9(3):237–247
SR Ahmad, AA Bakar, and MR Yaakub (2015). “Metaheuristic algorithms for feature selection in sentiment analysis,” In 2015 Science and Information Conference (SAI), pp. 222–226: IEEE
M Alweshah, E Ramadan, MH Ryalat, M Almi’ani, and AI Hammouri (2020). “Water Evaporation Algorithm With Probabilistic Neural Network For Solving Classification Problems,” Jordanian Journal of Computers and Information Technology (JJCIT), vol. 6, no. 01
Gokalp O, Tasci E, Ugur A (2020) A novel wrapper feature selection algorithm based on iterated greedy metaheuristic for sentiment classification. Expert Syst Appl 146:113176
Alweshah M, Abdullah S (2015) Hybridizing firefly algorithms with a probabilistic neural network for solving classification problems. Appl Soft Comput 35:513–524
Alweshah M, Alzubi OA, Alzubi JA, Alaqeel S (2016) Solving attribute reduction problem using wrapper genetic programming. International Journal of Computer Science and Network Security (IJCSNS) 16(5):77
Alweshah M (2019) Construction biogeography-based optimization algorithm for solving classification problems. Neural Comput & Applic 31(10):5679–5688
X-S Yang (2010). Engineering optimization: an introduction with metaheuristic applications. Wiley
Alweshah M, Hammouri AI, Tedmori S (2017) Biogeography-based optimisation for data classification problems. International Journal of Data Mining, Modelling and Management 9(2):142–162
Sayed GI, Tharwat A, Hassanien AE (2019) Chaotic dragonfly algorithm: an improved metaheuristic algorithm for feature selection. Appl Intell 49(1):188–205
Aghdam MH, Ghasem-Aghaee N, Basiri ME (2009) Text feature selection using ant colony optimization. Expert Syst Appl 36(3):6843–6853
Sakri SB, Rashid NBA, Zain ZM (2018) Particle swarm optimization feature selection for breast cancer recurrence prediction. IEEE Access 6:29637–29647
Shunmugapriya P, Kanmani S (2017) A hybrid algorithm using ant and bee colony optimization for feature selection and classification (AC-ABC hybrid). Swarm and Evolutionary Computation 36:27–36
Mafarja M, Aljarah I, Heidari AA, Hammouri AI, Faris H, al-Zoubi A’M, Mirjalili S (2018) Evolutionary population dynamics and grasshopper optimization approaches for feature selection problems. Knowl-Based Syst 145:25–45
Sayed GI, Hassanien AE, Azar AT (2019) Feature selection via a novel chaotic crow search algorithm. Neural Comput & Applic 31(1):171–188
Taradeh M, Mafarja M, Heidari AA, Faris H, Aljarah I, Mirjalili S, Fujita H (2019) An evolutionary gravitational search-based feature selection. Inf Sci 497:219–239
Lin K-C, Zhang K-Y, Huang Y-H, Hung JC, Yen N (2016) Feature selection based on an improved cat swarm optimization algorithm for big data classification. J Supercomput 72(8):3210–3221
Diao R, Shen Q (2012) Feature selection with harmony search. IEEE Transactions on Systems, Man, and Cybernetics, Part B (Cybernetics) 42(6):1509–1523
D Rodrigues et al. (2013). “BCS: A binary cuckoo search algorithm for feature selection,” In 2013 IEEE International Symposium on Circuits and Systems (ISCAS), pp. 465–468: IEEE
Alweshah M, Al-Daradkeh A, Al-Betar MA, Almomani A, Oqeili S (2019) $$\beta $$-hill climbing algorithm with probabilistic neural network for classification problems. J Ambient Intell Humaniz Comput:1–12
Al Nsour H, Alweshah M, Hammouri AI, Al Ofeishat H, Mirjalili S (2019) A hybrid Grey wolf Optimiser algorithm for solving time series classification problems. J Intell Syst
Mohammed Alweshah, Moad Abu Qadoura, Abdelaziz I. Hammouri, Mohd Sanusi Azmi, and S Alkhalaileh (2019). “Flower Pollination Algorithm for solving classification problems,” International Journal of Advances in Soft Computing and Its Applications, vol. In Press, pp. 1–13
Wang G-G, Deb S, Cui Z (2019) Monarch butterfly optimization. Neural Comput & Applic 31(7):1995–2014
Alweshah M, Al Khalaileh S, Gupta BB, Almomani A, Hammouri AI, Al-Betar MA (2020) The monarch butterfly optimization algorithm for solving feature selection problems. Neural Comput & Applic:1–15
Wang G-G, Deb S, Coelho LDS (2018) Earthworm optimisation algorithm: a bio-inspired metaheuristic algorithm for global optimisation problems. International Journal of Bio-Inspired Computation 12(1):1–22
Sayed GI, Khoriba G, Haggag MH (2018) A novel chaotic salp swarm algorithm for global optimization and feature selection. Appl Intell 48(10):3462–3481
Chatterjee B, Bhattacharyya T, Ghosh KK, Singh PK, Geem ZW, Sarkar R (2020) Late acceptance hill climbing based social ski driver algorithm for feature selection. IEEE Access 8:75393–75408
M Ghosh, S Malakar, S Bhowmik, R Sarkar, and M Nasipuri (2019). “Feature Selection for Handwritten Word Recognition Using Memetic Algorithm,” In Advances in Intelligent Computing: Springer, pp. 103–124
HM Zawbaa, E Emary, and B Parv (2015). “Feature selection based on antlion optimization algorithm,” in Complex Systems (WCCS), 2015 Third World Conference on, pp. 1–7: IEEE
C Shivalingegowda and P Jayasree (2020). “Hybrid gravitational search algorithm based model for optimizing coverage and connectivity in wireless sensor networks,” J Ambient Intell Humaniz Comput, pp. 1–14
Abdel-Basset M, El-Shahat D, El-henawy I, de Albuquerque VHC, Mirjalili S (2020) A new fusion of grey wolf optimizer algorithm with a two-phase mutation for feature selection. Expert Syst Appl 139:112824
S Wu (2015). “Comparative analysis of particle swarm optimization algorithms for text feature selection,”
Zawbaa HM, Emary E, Grosan C (2016) Feature selection via chaotic antlion optimization. PloS one 11(3):e0150652
J Wang, B Xue, X Gao, and M Zhang (2016). “A differential evolution approach to feature selection and instance selection,” In Pacific Rim International Conference on Artificial Intelligence, pp. 588–602: Springer
Shahbeig S, SADJAD K, SADEGHI M (2016) Feature selection from Iron direct reduction data based on binary differential evolution optimization. Bulletin de la Société Royale des Sciences de Liège 85:114–122
Malakar S, Ghosh M, Bhowmik S, Sarkar R, Nasipuri M (2019) A GA based hierarchical feature selection approach for handwritten word recognition. Neural Comput & Applic:1–20
R Saidi, W Bouaguel, and N Essoussi (2019). “Hybrid Feature Selection Method Based on the Genetic Algorithm and Pearson Correlation Coefficient,” In Machine Learning Paradigms: Theory and Application: Springer, pp. 3–24
Emary E, Zawbaa HM, Hassanien AE (2016) Binary ant lion approaches for feature selection. Neurocomputing 213:54–65
ME Basiri and S Nemati (2009). “A novel hybrid ACO-GA algorithm for text feature selection,” In Evolutionary Computation, 2009. CEC’09. IEEE Congress on, pp. 2561–2568: IEEE
Jona J, Nagaveni N (2014) Ant-cuckoo colony optimization for feature selection in digital mammogram. Pakistan journal of biological sciences: PJBS 17(2):266–271
Mafarja M, Abdullah S (2013) Investigating memetic algorithm in solving rough set attribute reduction. Int J Comput Appl Technol 48(3):195–202
Zorarpacı E, Özel SA (2016) A hybrid approach of differential evolution and artificial bee colony for feature selection. Expert Syst Appl 62:91–103
Zhang Y, Gong D-w, Cheng J (2015) Multi-objective particle swarm optimization approach for cost-based feature selection in classification. IEEE/ACM transactions on computational biology and bioinformatics 14(1):64–75
Zhang Y, Gong D, Hu Y, Zhang W (2015) Feature selection algorithm based on bare bones particle swarm optimization. Neurocomputing 148:150–157
Mafarja MM, Mirjalili S (2017) Hybrid whale optimization algorithm with simulated annealing for feature selection. Neurocomputing 260:302–312
Mafarja M, Aljarah I, Faris H, Hammouri AI, Ala’M A-Z, Mirjalili S (2019) Binary grasshopper optimisation algorithm approaches for feature selection problems. Expert Syst Appl 117:267–286
Faris H, Mafarja MM, Heidari AA, Aljarah I, al-Zoubi A’M, Mirjalili S, Fujita H (2018) An efficient binary salp swarm algorithm with crossover scheme for feature selection problems. Knowl-Based Syst 154:43–67
Alweshah M, Alkhalaileh S, Albashish D, Mafarja M, Bsoul Q, Dorgham O (2020) A hybrid mine blast algorithm for feature selection problems. Soft Comput:1–18
Feng Y, Yang J, Wu C, Lu M, Zhao X-J (2018) Solving 0–1 knapsack problems by chaotic monarch butterfly optimization algorithm with Gaussian mutation. Memetic Computing 10(2):135–150
S Chakrabarty, AK Pal, N Dey, D Das, and S Acharjee (2014). “Foliage area computation using Monarch butterfly algorithm,” In Non Conventional Energy (ICONCE), 2014 1st International Conference on, pp. 249–253: IEEE
Ghanem WA, Jantan A (2018) Hybridizing artificial bee colony with monarch butterfly optimization for numerical optimization problems. Neural Comput & Applic 30(1):163–181
Devikanniga D, Raj RJS (2018) Classification of osteoporosis by artificial neural network based on monarch butterfly optimisation algorithm. Healthcare technology letters 5(2):70–75
I Strumberger, M Sarac, D Markovic, and N Bacanin (2018). “Hybridized Monarch Butterfly Algorithm for Global Optimization Problems,” International Journal of Computers, vol. 3
Faris H, Aljarah I, Mirjalili S (2018) Improved monarch butterfly optimization for unconstrained global search and neural network training. Appl Intell 48(2):445–464
Yazdani S, Hadavandi E (2018) LMBO-DE: a linearized monarch butterfly optimization algorithm improved with differential evolution. Soft Comput:1–15
I Stromberger, E Tuba, N Bacanin, M Beko, and M Tuba (2018). “Monarch butterfly optimization algorithm for localization in wireless sensor networks,” In Radioelektronika (RADIOELEKTRONIKA), 2018 28th International Conference, pp. 1–6: IEEE
G-G Wang, X Zhao, and S Deb (2015). “A novel monarch butterfly optimization with greedy strategy and self-adaptive,” In Soft Computing and Machine Intelligence (ISCMI), 2015 Second International Conference on, pp. 45–50: IEEE
Feng Y, Wang G-G, Li W, Li N (2018) Multi-strategy monarch butterfly optimization algorithm for discounted {0-1} knapsack problem. Neural Comput & Applic 30(10):3019–3036
G-G Wang, X Zhao, and S Deb (2015). “A novel monarch butterfly optimization with greedy strategy and self-adaptive,” in 2015 Second International Conference on Soft Computing and Machine Intelligence (ISCMI), pp. 45–50: IEEE
Wang G-G, Deb S, Zhao X, Cui Z (2018) A new monarch butterfly optimization with an improved crossover operator. Oper Res 18(3):731–755
Brown CT, Liebovitch LS, Glendon R (2007) Lévy flights in Dobe Ju/‘hoansi foraging patterns. Hum Ecol 35(1):129–138
Pavlyukevich I (2007) Lévy flights, non-local search and simulated annealing. J Comput Phys 226(2):1830–1844
Reynolds AM, Frye MA (2007) Free-flight odor tracking in Drosophila is consistent with an optimal intermittent scale-free search. PloS one 2(4):e354
Barthelemy P, Bertolotti J, Wiersma DS (2008) A Lévy flight for light. Nature 453(7194):495
Shlesinger MF (2006) Mathematical physics: Search research. Nature 443(7109):281
P Lévy (1954). Théorie de l’addition des variables aléatoires. Gauthier-Villars
He X, Huang J, Rao Y, Gao L (2016) Chaotic teaching-learning-based optimization with Lévy flight for global numerical optimization. Computational intelligence and neuroscience 2016:43
Gandomi AH, Yang X-S, Alavi AH (2013) Cuckoo search algorithm: a metaheuristic approach to solve structural optimization problems. Eng Comput 29(1):17–35
Jensi R, Jiji GW (2016) An enhanced particle swarm optimization with levy flight for global optimization. Appl Soft Comput 43:248–261
H Liu and H Motoda (2007). Computational methods of feature selection. CRC Press
He J, Yao X (2001) Drift analysis and average time complexity of evolutionary algorithms. Artif Intell 127(1):57–85
Acknowledgements
The work published in this paper was funded by Al-Balqa Applied University, The Deanship of Scientific Research, Al-Salt, Jordan.
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Alweshah, M. Solving feature selection problems by combining mutation and crossover operations with the monarch butterfly optimization algorithm. Appl Intell 51, 4058–4081 (2021). https://doi.org/10.1007/s10489-020-01981-0
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-020-01981-0