Abstract
High dimensionality of data represents a major problem that affects the accuracy of the classification. This problem related with classification is mainly resulted from the availability of irrelevant features. Feature selection represents a solution to a problem by selecting the most informative features and discard the irrelevant features. Generalized normal distribution optimization (GNDO) represents a newly developed optimization that confirmed its outperformance in comparison with well-known optimization algorithms on parameter extraction for photovoltaic models. As an optimization algorithm, however, GNDO suffers from degraded performance when dealing with a problem with a high dimensionality. The main problems of GNDO include exploitation problem by falling into local optima problem. Also, GNDO has solutions diversity problem when it deals with data with high dimensionality. To alleviate the drawbacks of this algorithm and solve feature selection problems, a local search algorithm (LSA) is used. The new algorithm is called dynamic generalized normal distribution optimization (DGNDO), which includes the following main improvements to GNDO: it can improve the best solution to solve the local optima problem, it can improve solution diversity by improving the randomly selected solution, and it can improve both exploration and exploitation combined. To confirm the outperformance and efficiency of the new DGNDO algorithm, DGNDO algorithm is applied on 20 benchmarked datasets from UCI repository of data. In addition, DGNDO algorithm results are compared with seven well-known optimization algorithms using number of evaluation metrics including classification, accuracy, fitness, the number of selected features, statistical results using Wilcoxon test and convergence curves. The obtained results reveal the superiority of DGNDO algorithm over all other competing algorithms.
Similar content being viewed by others
References
Tubishat M, Idris N, Shuib L, Abushariah MA, Mirjalili S (2020) Improved Salp swarm algorithm based on opposition based learning and novel local search algorithm for feature selection. Expert Syst Appl 145:113122
Sihwail R, Omar K, Ariffin KAZ, Tubishat M (2020) Improved harris hawks optimization using elite opposition-based learning and novel search mechanism for feature selection. IEEE Access 8:121127–121145
Abdel-Basset M, Ding W, El-Shahat D (2021) A hybrid Harris Hawks optimization algorithm with simulated annealing for feature selection. Artif Intell Rev 54:593–637
Agrawal R, Kaur B, Sharma S (2020) Quantum based whale optimization algorithm for wrapper feature selection. Appl Soft Comput 89:106092
Tubishat M, Abushariah MA, Idris N, Aljarah I (2019) Improved whale optimization algorithm for feature selection in Arabic sentiment analysis. Appl Intell 49:1688–1707
Tubishat M, Ja’afar S, Alswaitti M, Mirjalili S, Idris N, Ismail MA et al (2020) Dynamic salp swarm algorithm for feature selection. Expert Syst Appl 164:113873
Tubishat M, Alswaitti M, Mirjalili S, Al-Garadi MA, Rana TA (2020) Dynamic butterfly optimization algorithm for feature selection. IEEE Access 8:194303–194314
Elgamal ZM, Yasin NBM, Tubishat M, Alswaitti M, Mirjalili S (2020) An improved Harris Hawks optimization algorithm with simulated annealing for feature selection in the medical field. IEEE Access 8:186638–186652
Alweshah M, Al Khalaileh S, Gupta BB, Almomani A, Hammouri AI, Al-Betar MA (2020) The monarch butterfly optimization algorithm for solving feature selection problems. Neural Comput Appl. https://doi.org/10.1007/s00521-020-05210-0
Neggaz N, Houssein EH, Hussain K (2020) An efficient henry gas solubility optimization for feature selection. Expert Syst Appl 152:113364
Abdel-Basset M, El-Shahat D, El-henawy I, de Albuquerque VHC, Mirjalili S (2020) A new fusion of grey wolf optimizer algorithm with a two-phase mutation for feature selection. Expert Syst Appl 139:112824
Guha R, Ghosh M, Mutsuddi S, Sarkar R, Mirjalili S (2020) Embedded chaotic whale survival algorithm for filter–wrapper feature selection. Soft Comput 24:12821–12843
Faris H, Heidari AA, Ala’M A-Z, Mafarja M, Aljarah I, Eshtay M et al (2020) Time-varying hierarchical chains of salps with random weight networks for feature selection. Expert Syst Appl 140:1128
Amini F, Hu G (2021) A two-layer feature selection method using genetic algorithm and elastic net. Expert Syst Appl 166:114072
Alweshah M, Alkhalaileh S, Albashish D, Mafarja M, Bsoul Q, Dorgham O (2021) A hybrid mine blast algorithm for feature selection problems. Soft Comput 25:517–534
Kılıç F, Kaya Y, Yildirim S (2021) A novel multi population based particle swarm optimization for feature selection. Knowl-Based Syst 219:106894
Tubishat M, S. Ja’afar, N. Idris, M. A. Al-Betar, M. Alswaitti, H. Jarrah, et al (2021) Improved sine cosine algorithm with simulated annealing and singer chaotic map for Hadith classification. Neural Comput Appl 34(2):1385–1406
Khurmaa RA, Aljarah I, Sharieh A (2020) An intelligent feature selection approach based on moth flame optimization for medical diagnosis. Neural Comput Appl 33(12):7165–7204
Ahmed S, Ghosh KK, Garcia-Hernandez L, Abraham A, Sarkar R (2021) Improved coral reefs optimization with adaptive $$\beta $$ β-hill climbing for feature selection. Neural Comput Appl 33:6467–6486
Divya R, Kumari RSS (2021) Genetic algorithm with logistic regression feature selection for Alzheimer’s disease classification. Neural Comput Appl 33(14):8435–8444
Wolpert DH, Macready WG (1997) No free lunch theorems for optimization. IEEE Trans Evol Comput 1:67–82
Rao H, Shi X, Rodrigue AK, Feng J, Xia Y, Elhoseny M et al (2019) Feature selection based on artificial bee colony and gradient boosting decision tree. Appl Soft Comput 74:634–642
Hu P, Pan J-S, Chu S-C (2020) Improved binary grey wolf optimizer and its application for feature selection. Knowl-Based Syst 195:105746
Alazzam H, Sharieh A, Sabri KE (2020) A feature selection algorithm for intrusion detection system based on pigeon inspired optimizer. Expert Syst Appl 148:113249
Zhang X, Xu Y, Yu C, Heidari AA, Li S, Chen H et al (2020) Gaussian mutational chaotic fruit fly-built optimization and feature selection. Expert Syst Appl 141:112976
Zhang Y, Jin Z, Mirjalili S (2020) Generalized normal distribution optimization and its applications in parameter extraction of photovoltaic models. Energy Convers Manage 224:113301
Abdel-Basset M, Manogaran G, El-Shahat D, Mirjalili S (2018) A hybrid whale optimization algorithm based on local search strategy for the permutation flow shop scheduling problem. Futur Gener Comput Syst 85:129–145
Tubishat M, Idris N, Abushariah M (2021) Explicit aspects extraction in sentiment analysis using optimal rules combination. Futur Gener Comput Syst 114:448–480
Toksari MD (2016) A hybrid algorithm of Ant Colony Optimization (ACO) and Iterated Local Search (ILS) for estimating electricity domestic consumption: case of Turkey. Int J Electr Power Energy Syst 78:776–782
Mavrovouniotis M, Müller FM, Yang S (2016) Ant colony optimization with local search for dynamic traveling salesman problems. IEEE Trans Cybern 47:1743–1756
Mafarja MM, Mirjalili S (2017) Hybrid whale optimization algorithm with simulated annealing for feature selection. Neurocomputing 260:302–312
Sakamoto S, Ozera K, Ikeda M, Barolli L (2018) Implementation of intelligent hybrid systems for node placement problem in WMNs considering particle swarm optimization, hill climbing and simulated annealing. Mob Netw Appl 23:27–33
Yan C, Ma J, Luo H, Patel A (2019) Hybrid binary coral reefs optimization algorithm with simulated annealing for feature selection in high-dimensional biomedical datasets. Chemom Intell Lab Syst 184:102–111
Shehab M, Khader AT, Al-Betar MA, Abualigah LM (2017) Hybridizing cuckoo search algorithm with hill climbing for numerical optimization problems. In: 2017 8th International conference on information technology (ICIT), pp 36–43
Zhao F, Qin S, Zhang Y, Ma W, Zhang C, Song H (2019) A hybrid biogeography-based optimization with variable neighborhood search mechanism for no-wait flow shop scheduling problem. Expert Syst Appl 126:321–339
Riahi V, Kazemi M (2018) A new hybrid ant colony algorithm for scheduling of no-wait flowshop. Oper Res Int J 18:55–74
Asadzadeh L (2015) A local search genetic algorithm for the job shop scheduling problem with intelligent agents. Comput Ind Eng 85:376–383
Nekkaa M, Boughaci D (2016) Hybrid harmony search combined with stochastic local search for feature selection. Neural Process Lett 44:199–220
Ou T-C, Su W-F, Liu X-Z, Huang S-J, Tai T-Y (2016) A modified bird-mating optimization with hill-climbing for connection decisions of transformers. Energies 9:671
Pei J, Liu X, Fan W, Pardalos PM, Lu S (2019) A hybrid BA-VNS algorithm for coordinated serial-batching scheduling with deteriorating jobs, financial budget, and resource constraint in multiple manufacturers. Omega 82:55–69
Hussain K, Neggaz N, Zhu W, Houssein EH (2021) An efficient hybrid sine-cosine Harris hawks optimization for low and high-dimensional feature selection. Expert Syst Appl 176:114778
Too J, Mirjalili S (2021) A hyper learning binary dragonfly algorithm for feature selection: a COVID-19 case study. Knowl-Based Syst 212:106553
Aljarah I, Mafarja M, Heidari AA, Faris H, Zhang Y, Mirjalili S (2018) Asynchronous accelerating multi-leader salp chains for feature selection. Appl Soft Comput 71:964–979
Mafarja M, Aljarah I, Heidari AA, Hammouri AI, Faris H, Ala’M A-Z et al (2018) Evolutionary population dynamics and grasshopper optimization approaches for feature selection problems. Knowl-Based Syst 145:25–45
Hegazy AE, Makhlouf M, El-Tawel GS (2020) Improved salp swarm algorithm for feature selection. J King Saud Univ-Comput Inf Sci 32:335–344
Hegazy AE, Makhlouf M, El-Tawel GS (2019) Feature selection using chaotic salp swarm algorithm for data classification. Arab J Sci Eng 44:3801–3816
Arora S, Anand P (2019) Binary butterfly optimization approaches for feature selection. Expert Syst Appl 116:147–160
Saremi S, Mirjalili S, Lewis A (2017) Grasshopper optimisation algorithm: theory and application. Adv Eng Softw 105:30–47
Thaher T, Heidari AA, Mafarja M, Dong JS, Mirjalili S (2020) Binary Harris Hawks optimizer for high-dimensional, low sample size feature selection. In: Mirjalili S, Faris H, Aljarah I (eds) Evolutionary machine learning techniques. Springer, Cham, pp 251–272
Emine B, Ülker E (2020) An efficient binary social spider algorithm for feature selection problem. Expert Syst Appl 146:113185
Arora S, Singh S (2019) Butterfly optimization algorithm: a novel approach for global optimization. Soft Comput 23:715–734
Mirjalili S (2015) The ant lion optimizer. Adv Eng Softw 83:80–98
Mirjalili S (2016) SCA: a sine cosine algorithm for solving optimization problems. Knowl-Based Syst 96:120–133
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of interest
The authors declare that they have no conflict of interest.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Tubishat, M., Rawshdeh, Z., Jarrah, H. et al. Dynamic generalized normal distribution optimization for feature selection. Neural Comput & Applic 34, 17355–17370 (2022). https://doi.org/10.1007/s00521-022-07398-9
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s00521-022-07398-9