Skip to main content
Log in

Dynamic generalized normal distribution optimization for feature selection

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

High dimensionality of data represents a major problem that affects the accuracy of the classification. This problem related with classification is mainly resulted from the availability of irrelevant features. Feature selection represents a solution to a problem by selecting the most informative features and discard the irrelevant features. Generalized normal distribution optimization (GNDO) represents a newly developed optimization that confirmed its outperformance in comparison with well-known optimization algorithms on parameter extraction for photovoltaic models. As an optimization algorithm, however, GNDO suffers from degraded performance when dealing with a problem with a high dimensionality. The main problems of GNDO include exploitation problem by falling into local optima problem. Also, GNDO has solutions diversity problem when it deals with data with high dimensionality. To alleviate the drawbacks of this algorithm and solve feature selection problems, a local search algorithm (LSA) is used. The new algorithm is called dynamic generalized normal distribution optimization (DGNDO), which includes the following main improvements to GNDO: it can improve the best solution to solve the local optima problem, it can improve solution diversity by improving the randomly selected solution, and it can improve both exploration and exploitation combined. To confirm the outperformance and efficiency of the new DGNDO algorithm, DGNDO algorithm is applied on 20 benchmarked datasets from UCI repository of data. In addition, DGNDO algorithm results are compared with seven well-known optimization algorithms using number of evaluation metrics including classification, accuracy, fitness, the number of selected features, statistical results using Wilcoxon test and convergence curves. The obtained results reveal the superiority of DGNDO algorithm over all other competing algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

References

  1. Tubishat M, Idris N, Shuib L, Abushariah MA, Mirjalili S (2020) Improved Salp swarm algorithm based on opposition based learning and novel local search algorithm for feature selection. Expert Syst Appl 145:113122

    Google Scholar 

  2. Sihwail R, Omar K, Ariffin KAZ, Tubishat M (2020) Improved harris hawks optimization using elite opposition-based learning and novel search mechanism for feature selection. IEEE Access 8:121127–121145

    Google Scholar 

  3. Abdel-Basset M, Ding W, El-Shahat D (2021) A hybrid Harris Hawks optimization algorithm with simulated annealing for feature selection. Artif Intell Rev 54:593–637

    Google Scholar 

  4. Agrawal R, Kaur B, Sharma S (2020) Quantum based whale optimization algorithm for wrapper feature selection. Appl Soft Comput 89:106092

    Google Scholar 

  5. Tubishat M, Abushariah MA, Idris N, Aljarah I (2019) Improved whale optimization algorithm for feature selection in Arabic sentiment analysis. Appl Intell 49:1688–1707

    Google Scholar 

  6. Tubishat M, Ja’afar S, Alswaitti M, Mirjalili S, Idris N, Ismail MA et al (2020) Dynamic salp swarm algorithm for feature selection. Expert Syst Appl 164:113873

    Google Scholar 

  7. Tubishat M, Alswaitti M, Mirjalili S, Al-Garadi MA, Rana TA (2020) Dynamic butterfly optimization algorithm for feature selection. IEEE Access 8:194303–194314

    Google Scholar 

  8. Elgamal ZM, Yasin NBM, Tubishat M, Alswaitti M, Mirjalili S (2020) An improved Harris Hawks optimization algorithm with simulated annealing for feature selection in the medical field. IEEE Access 8:186638–186652

    Google Scholar 

  9. Alweshah M, Al Khalaileh S, Gupta BB, Almomani A, Hammouri AI, Al-Betar MA (2020) The monarch butterfly optimization algorithm for solving feature selection problems. Neural Comput Appl. https://doi.org/10.1007/s00521-020-05210-0

    Article  Google Scholar 

  10. Neggaz N, Houssein EH, Hussain K (2020) An efficient henry gas solubility optimization for feature selection. Expert Syst Appl 152:113364

    Google Scholar 

  11. Abdel-Basset M, El-Shahat D, El-henawy I, de Albuquerque VHC, Mirjalili S (2020) A new fusion of grey wolf optimizer algorithm with a two-phase mutation for feature selection. Expert Syst Appl 139:112824

    Google Scholar 

  12. Guha R, Ghosh M, Mutsuddi S, Sarkar R, Mirjalili S (2020) Embedded chaotic whale survival algorithm for filter–wrapper feature selection. Soft Comput 24:12821–12843

    Google Scholar 

  13. Faris H, Heidari AA, Ala’M A-Z, Mafarja M, Aljarah I, Eshtay M et al (2020) Time-varying hierarchical chains of salps with random weight networks for feature selection. Expert Syst Appl 140:1128

    Google Scholar 

  14. Amini F, Hu G (2021) A two-layer feature selection method using genetic algorithm and elastic net. Expert Syst Appl 166:114072

    Google Scholar 

  15. Alweshah M, Alkhalaileh S, Albashish D, Mafarja M, Bsoul Q, Dorgham O (2021) A hybrid mine blast algorithm for feature selection problems. Soft Comput 25:517–534

    Google Scholar 

  16. Kılıç F, Kaya Y, Yildirim S (2021) A novel multi population based particle swarm optimization for feature selection. Knowl-Based Syst 219:106894

    Google Scholar 

  17. Tubishat M, S. Ja’afar, N. Idris, M. A. Al-Betar, M. Alswaitti, H. Jarrah, et al (2021) Improved sine cosine algorithm with simulated annealing and singer chaotic map for Hadith classification. Neural Comput Appl 34(2):1385–1406

    Google Scholar 

  18. Khurmaa RA, Aljarah I, Sharieh A (2020) An intelligent feature selection approach based on moth flame optimization for medical diagnosis. Neural Comput Appl 33(12):7165–7204

    Google Scholar 

  19. Ahmed S, Ghosh KK, Garcia-Hernandez L, Abraham A, Sarkar R (2021) Improved coral reefs optimization with adaptive $$\beta $$ β-hill climbing for feature selection. Neural Comput Appl 33:6467–6486

    Google Scholar 

  20. Divya R, Kumari RSS (2021) Genetic algorithm with logistic regression feature selection for Alzheimer’s disease classification. Neural Comput Appl 33(14):8435–8444

    Google Scholar 

  21. Wolpert DH, Macready WG (1997) No free lunch theorems for optimization. IEEE Trans Evol Comput 1:67–82

    Google Scholar 

  22. Rao H, Shi X, Rodrigue AK, Feng J, Xia Y, Elhoseny M et al (2019) Feature selection based on artificial bee colony and gradient boosting decision tree. Appl Soft Comput 74:634–642

    Google Scholar 

  23. Hu P, Pan J-S, Chu S-C (2020) Improved binary grey wolf optimizer and its application for feature selection. Knowl-Based Syst 195:105746

    Google Scholar 

  24. Alazzam H, Sharieh A, Sabri KE (2020) A feature selection algorithm for intrusion detection system based on pigeon inspired optimizer. Expert Syst Appl 148:113249

    Google Scholar 

  25. Zhang X, Xu Y, Yu C, Heidari AA, Li S, Chen H et al (2020) Gaussian mutational chaotic fruit fly-built optimization and feature selection. Expert Syst Appl 141:112976

    Google Scholar 

  26. Zhang Y, Jin Z, Mirjalili S (2020) Generalized normal distribution optimization and its applications in parameter extraction of photovoltaic models. Energy Convers Manage 224:113301

    Google Scholar 

  27. Abdel-Basset M, Manogaran G, El-Shahat D, Mirjalili S (2018) A hybrid whale optimization algorithm based on local search strategy for the permutation flow shop scheduling problem. Futur Gener Comput Syst 85:129–145

    Google Scholar 

  28. Tubishat M, Idris N, Abushariah M (2021) Explicit aspects extraction in sentiment analysis using optimal rules combination. Futur Gener Comput Syst 114:448–480

    Google Scholar 

  29. Toksari MD (2016) A hybrid algorithm of Ant Colony Optimization (ACO) and Iterated Local Search (ILS) for estimating electricity domestic consumption: case of Turkey. Int J Electr Power Energy Syst 78:776–782

    Google Scholar 

  30. Mavrovouniotis M, Müller FM, Yang S (2016) Ant colony optimization with local search for dynamic traveling salesman problems. IEEE Trans Cybern 47:1743–1756

    Google Scholar 

  31. Mafarja MM, Mirjalili S (2017) Hybrid whale optimization algorithm with simulated annealing for feature selection. Neurocomputing 260:302–312

    Google Scholar 

  32. Sakamoto S, Ozera K, Ikeda M, Barolli L (2018) Implementation of intelligent hybrid systems for node placement problem in WMNs considering particle swarm optimization, hill climbing and simulated annealing. Mob Netw Appl 23:27–33

    Google Scholar 

  33. Yan C, Ma J, Luo H, Patel A (2019) Hybrid binary coral reefs optimization algorithm with simulated annealing for feature selection in high-dimensional biomedical datasets. Chemom Intell Lab Syst 184:102–111

    Google Scholar 

  34. Shehab M, Khader AT, Al-Betar MA, Abualigah LM (2017) Hybridizing cuckoo search algorithm with hill climbing for numerical optimization problems. In: 2017 8th International conference on information technology (ICIT), pp 36–43

  35. Zhao F, Qin S, Zhang Y, Ma W, Zhang C, Song H (2019) A hybrid biogeography-based optimization with variable neighborhood search mechanism for no-wait flow shop scheduling problem. Expert Syst Appl 126:321–339

    Google Scholar 

  36. Riahi V, Kazemi M (2018) A new hybrid ant colony algorithm for scheduling of no-wait flowshop. Oper Res Int J 18:55–74

    Google Scholar 

  37. Asadzadeh L (2015) A local search genetic algorithm for the job shop scheduling problem with intelligent agents. Comput Ind Eng 85:376–383

    Google Scholar 

  38. Nekkaa M, Boughaci D (2016) Hybrid harmony search combined with stochastic local search for feature selection. Neural Process Lett 44:199–220

    Google Scholar 

  39. Ou T-C, Su W-F, Liu X-Z, Huang S-J, Tai T-Y (2016) A modified bird-mating optimization with hill-climbing for connection decisions of transformers. Energies 9:671

    Google Scholar 

  40. Pei J, Liu X, Fan W, Pardalos PM, Lu S (2019) A hybrid BA-VNS algorithm for coordinated serial-batching scheduling with deteriorating jobs, financial budget, and resource constraint in multiple manufacturers. Omega 82:55–69

    Google Scholar 

  41. Hussain K, Neggaz N, Zhu W, Houssein EH (2021) An efficient hybrid sine-cosine Harris hawks optimization for low and high-dimensional feature selection. Expert Syst Appl 176:114778

    Google Scholar 

  42. Too J, Mirjalili S (2021) A hyper learning binary dragonfly algorithm for feature selection: a COVID-19 case study. Knowl-Based Syst 212:106553

    Google Scholar 

  43. Aljarah I, Mafarja M, Heidari AA, Faris H, Zhang Y, Mirjalili S (2018) Asynchronous accelerating multi-leader salp chains for feature selection. Appl Soft Comput 71:964–979

    Google Scholar 

  44. Mafarja M, Aljarah I, Heidari AA, Hammouri AI, Faris H, Ala’M A-Z et al (2018) Evolutionary population dynamics and grasshopper optimization approaches for feature selection problems. Knowl-Based Syst 145:25–45

    Google Scholar 

  45. Hegazy AE, Makhlouf M, El-Tawel GS (2020) Improved salp swarm algorithm for feature selection. J King Saud Univ-Comput Inf Sci 32:335–344

    Google Scholar 

  46. Hegazy AE, Makhlouf M, El-Tawel GS (2019) Feature selection using chaotic salp swarm algorithm for data classification. Arab J Sci Eng 44:3801–3816

    Google Scholar 

  47. Arora S, Anand P (2019) Binary butterfly optimization approaches for feature selection. Expert Syst Appl 116:147–160

    Google Scholar 

  48. Saremi S, Mirjalili S, Lewis A (2017) Grasshopper optimisation algorithm: theory and application. Adv Eng Softw 105:30–47

    Google Scholar 

  49. Thaher T, Heidari AA, Mafarja M, Dong JS, Mirjalili S (2020) Binary Harris Hawks optimizer for high-dimensional, low sample size feature selection. In: Mirjalili S, Faris H, Aljarah I (eds) Evolutionary machine learning techniques. Springer, Cham, pp 251–272

    Google Scholar 

  50. Emine B, Ülker E (2020) An efficient binary social spider algorithm for feature selection problem. Expert Syst Appl 146:113185

    Google Scholar 

  51. Arora S, Singh S (2019) Butterfly optimization algorithm: a novel approach for global optimization. Soft Comput 23:715–734

    Google Scholar 

  52. Mirjalili S (2015) The ant lion optimizer. Adv Eng Softw 83:80–98

    Google Scholar 

  53. Mirjalili S (2016) SCA: a sine cosine algorithm for solving optimization problems. Knowl-Based Syst 96:120–133

    Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Mohammad Tubishat.

Ethics declarations

Conflict of interest

The authors declare that they have no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Tubishat, M., Rawshdeh, Z., Jarrah, H. et al. Dynamic generalized normal distribution optimization for feature selection. Neural Comput & Applic 34, 17355–17370 (2022). https://doi.org/10.1007/s00521-022-07398-9

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-022-07398-9

Keywords

Navigation