Skip to main content
Log in

Enhanced feature selection technique using slime mould algorithm: a case study on chemical data

  • Original Article
  • Published:
Neural Computing and Applications Aims and scope Submit manuscript

Abstract

Feature selection techniques are considered one of the most important preprocessing steps, which has the most significant influence on the performance of data analysis and decision making. These FS techniques aim to achieve several objectives (such as reducing classification error and minimizing the number of features) at the same time to increase the classification rate. FS based on Metaheuristic (MH) is considered one of the most promising techniques to improve the classification process. This paper presents a modified method of the Slime mould algorithm depending on the Marine Predators Algorithm (MPA) operators as a local search strategy, which leads to increasing the convergence rate of the developed method, named SMAMPA and avoiding the attraction to local optima. The efficiency of SMAMPA is evaluated using twenty datasets and compared its results with the state-of-the-art FS methods. In addition, the applicability of SMAMPA to work with real-world problems is evaluated by using it as a quantitative structure-activity relationship (QSAR) model. The obtained results show the high ability of the developed SMAMPA method to reduce the dimension of the tested datasets by increasing the prediction rate. In addition, it provides results better than other FS techniques in terms of performance metrics.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data Availability

The datasets generated during and/or analysed during the current study are available in the UCI repository [61].

References

  1. Quiroz Juan C, Amit B, Dascalu Sergiu M, Lun Lau S (2017) Feature selection for activity recognition from smartphone accelerometer data. Intelli Autom Soft Comput 87:1–9

    Google Scholar 

  2. Han C, Zhou G, Zhou Y (2019) Binary symbiotic organism search algorithm for feature selection and analysis. IEEE Access 7:166833–166859

    Article  Google Scholar 

  3. Han J, Pei J, Kamber M (2011) Data mining: concepts and techniques. Elsevier, Amsterdam

    MATH  Google Scholar 

  4. Hancer E, Xue B, Karaboga D, Zhang M (2015) A binary abc algorithm based on advanced similarity scheme for feature selection. Appl Soft Comput 36:334–348

    Article  Google Scholar 

  5. Hua J, Tembe Waibhav D, Dougherty Edward R (2009) Performance of feature-selection methods in the classification of high-dimension data. Pattern Recogn 42(3):409–424

    Article  MATH  Google Scholar 

  6. Jingwei T, Seyedali M (2020) A hyper learning binary dragonfly algorithm for feature selection: a covid-19 case study. Knowl-Based Syst 87:106553

    Google Scholar 

  7. Zhong C, Chen Y, Jian P (2020) Feature selection based on a novel improved tree growth algorithm. Int J Comput Intell Syst 13(1):247–258

    Article  Google Scholar 

  8. Xue B, Zhang M, Browne WN (2014) Novel initialisation and updating mechanisms. particle swarm optimisation for feature selection in classification. Appl Soft Comput 18:261–276

    Article  Google Scholar 

  9. Tan F, Xuezheng F, Zhang Y, Bourgeois Anu G (2008) A genetic algorithm-based method for feature subset selection. Soft Comput 12(2):111–120

    Article  Google Scholar 

  10. Mustafa Serter U, Nihat Y, Onur I(2013) Feature selection method based on artificial bee colony algorithm and support vector machines for medical datasets classification. The Scientific World Journal 2013

  11. Selvakumar B, Muneeswaran K (2019) Firefly algorithm based feature selection for network intrusion detection. Computers Secur 81:148–155

    Article  Google Scholar 

  12. Emary E, Zawbaa Hossam M, Hassanien Aboul E (2016) Binary grey wolf optimization approaches for feature selection. Neurocomputing 172:371–381

    Article  Google Scholar 

  13. Sindhu R, Ngadiran R, Yacob YM, Zahri Nik Adilah H, Hariharan M (2017) Sine-cosine algorithm for feature selection with elitism strategy and new updating mechanism. Neural Comput Appl 28(10):2947–2958

    Article  Google Scholar 

  14. Faris H, Mafarja Majdi M, Heidari Ali A, Aljarah I, Ala’M A-Z, Mirjalili S, Fujita H (2018) An efficient binary salp swarm algorithm with crossover scheme for feature selection problems. Knowl-Based Syst 154:43–67

    Article  Google Scholar 

  15. Ewees Ahmed A, Aziz Mohamed AE, Hassanien Aboul E (2019) Chaotic multi-verse optimizer-based feature selection. Neural Comput Appl 31(4):991–1006

    Article  Google Scholar 

  16. Abualigah L, Diabat A, Mirjalili S, Elaziz MA, Gandomi AH (2021) The arithmetic optimization algorithm. Computer Methods Appl Mech Eng 376:113609

    Article  MATH  Google Scholar 

  17. Laith A, Ali D (2021) Advances in sine cosine algorithm: a comprehensive survey. Artif Intell Rev 25:1–42

    Google Scholar 

  18. Ewees Ahmed A, Al-qaness Mohammed AA, Abualigah L, Oliva D, Algamal ZY, Anter AM, Ibrahim RA, Ghoniem RM, Elaziz MA (2021) Boosting arithmetic optimization algorithm with genetic algorithm operators for feature selection: case study on cox proportional hazards model. Mathematics 9(18):2321

    Article  Google Scholar 

  19. Ibrahim Rehab A, Ewees Ahmed A, Oliva D, Elaziz MA, Songfeng L (2019) Improved salp swarm algorithm based on particle swarm optimization for feature selection. J Amb Intell Humanized Comput 10(8):3155–3169

    Article  Google Scholar 

  20. Zorarpaci E, Aycseozel S (2016) A hybrid approach of differential evolution and artificial bee colony for feature selection. Expert Syst Appl 62:91–103

    Article  Google Scholar 

  21. Arora S, Singh H, Sharma M, Sharma S, Anand P (2019) A new hybrid algorithm based on grey wolf optimization and crow search algorithm for unconstrained function optimization and feature selection. IEEE Access 7:26343–26361

    Article  Google Scholar 

  22. Abd Mohamed E, Elaziz Ahmed A, Diego Oliva E, Pengfei D, Shengwu X (2017) A hybrid method of sine cosine algorithm and differential evolution for feature selection. In International conference on neural information processing, 145–155 Springer,

  23. Elaziz MA, Ewees Ahmed A, Ibrahim RA, Songfeng L (2020) Opposition-based moth-flame optimization improved by differential evolution for feature selection. Math Computers Simul 168:48–75

    Article  MATH  Google Scholar 

  24. Neggaz N, Ewees Ahmed A, Elaziz MA, Mafarja M (2020) Boosting salp swarm algorithm by sine cosine algorithm and disrupt operator for feature selection. Expert Syst Appl 145:113103

    Article  Google Scholar 

  25. Laith A, Ali D (2020) A comprehensive survey of the grasshopper optimization algorithm: results, variants, and applications. Neural Comput Appl 25:1–24

    Google Scholar 

  26. Shimin L, Huiling C, Mingjing W, Asghar Heidari A, and Mirjalili S (2020) A new method for stochastic optimization. future generation computer systems, Slime mould algorithm

  27. Kumar C, Dharma Raj T, Premkumar M, Dhanesh Raj T (2020) A new stochastic slime mould optimization algorithm for the estimation of solar photovoltaic cell parameters. Optik 223:165277

    Article  Google Scholar 

  28. Chen Z, Liu W (2020) An efficient parameter adaptive support vector regression using k-means clustering and chaotic slime mould algorithm. IEEE Access 8:156851–156862

    Article  Google Scholar 

  29. Al-Qaness Mohammed AA, Hong F, Ewees Ahmed A, Dalia Y, Mohammed Abd E (2020) Improved anfis model for forecasting Wuhan city air quality and analysis covid-19 lockdown impacts on air quality. Environ Res 871:110607

    Google Scholar 

  30. Ali D (2020) The optimal synthesis of thinned concentric circular antenna arrays using slime mold algorithm. Electromagnetics 58:1–13

    Google Scholar 

  31. Sun K, Jia H, Li Y, Jiang Z (2021) Hybrid improved slime mould algorithm with adaptive \(\beta\) hill climbing for numerical optimization. J Intell Fuzzy Syst (Preprint) 14:1667–1679

    Article  Google Scholar 

  32. Ewees Ahmed A, Laith A, Dalia Y, Zakariya Yahya A, Al-Ganess Mohammed AA, Rehab Ali I, Mohamed Abd E (2021) Improved slime mould algorithm based on firefly algorithm for feature selection: a case study on qsar model. Eng Computers 69:1–15

    Google Scholar 

  33. Afshin F, Mohammad H, Seyedali M, Gandomi Amir H (2020) Marine predators algorithm: a nature-inspired metaheuristic. Expert Syst Appl 5:113377

    Google Scholar 

  34. Al-Qaness Mohammed AA, Saba Amal I, Elsheikh Ammar H, Elaziz MA, Ibrahim Rehab A, Songfeng L, Hemedan Ahmed A, Shanmugan S, Ewees Ahmed A (2020) Efficient artificial intelligence forecasting models for covid-19 outbreak in russia and brazil. Process Safety and Environmental Protection

  35. Al-Qaness Mohammed AA, Ewees Ahmed A, Fan H, Abualigah L, Elaziz MA (2020) Marine predators algorithm for forecasting confirmed cases of Covid-19 in Italy, USA, Iran and Korea. Int J Environ Res Publ Health 17(10):3520

    Article  Google Scholar 

  36. Elaziz MA, Ewees Ahmed A, Yousri D, Naji Husein S, Alwerfali Qamar A, Awad Songfeng L, Al-Qaness Mohammed AA (2020) An improved marine predators algorithm with fuzzy entropy for multi-level thresholding: Real world example of Covid-19 ct image segmentation. IEEE Access 8:125306–125330

    Article  Google Scholar 

  37. Sahlol Ahmed T, Yousri D, Ewees Ahmed A, Al-Qaness MAA, Damasevicius R, Elaziz MA (2020) Covid-19 image classification using deep features and fractional-order marine predators algorithm. Scientif Rep 10(1):1–15

    Google Scholar 

  38. Yousri D, Hasanien Hany M, Fathy A (2020) Parameters identification of solid oxide fuel cell for static and dynamic simulation using comprehensive learning dynamic multi-swarm marine predators algorithm. Energy Conver Manage 228:113692

    Article  Google Scholar 

  39. Elaziz MA, Shehabeldeen Taher A, Elsheikh Ammar H, Zhou J, Ewees Ahmed A, Al-qaness Mohammed AA (2020) Utilization of random vector functional link integrated with marine predators algorithm for tensile behavior prediction of dissimilar friction stir welded aluminum alloy joints. J Mater Res Technol 9(5):11370–11381

    Article  Google Scholar 

  40. Al-qaness MAA, Ewees AA, Fan H, Abualigah L, Elaziz MA (2022) Boosted anfis model using augmented marine predator algorithm with mutation operators for wind power forecasting. Appl Energy 314:118851

    Article  Google Scholar 

  41. Yousri D, Babu TS, Beshr E, Eteiba Magdy B, Allam D (2020) A robust strategy based on marine predators algorithm for large scale photovoltaic array reconfiguration to mitigate the partial shading effect on the performance of pv system. IEEE Access 8:112407–112426

    Article  Google Scholar 

  42. Sayed Safinaz A-F, Nabil E, Badr A (2016) A binary clonal flower pollination algorithm for feature selection. Pattern Recognit Lett 77:21–27

    Article  Google Scholar 

  43. Mafarja M, Mirjalili S (2018) Whale optimization approaches for wrapper feature selection. Appl Soft Comput 62:441–453

    Article  Google Scholar 

  44. Sayed GI, Hassanien AE, Azar AT (2019) Feature selection via a novel chaotic crow search algorithm. Neural Comput Appl 31(1):171–188

    Article  Google Scholar 

  45. Arora S, Anand P (2019) Binary butterfly optimization approaches for feature selection. Expert Syst Appl 116:147–160

    Article  Google Scholar 

  46. Jingwei T, Abdul Rahim A (2020) A new and fast rival genetic algorithm for feature selection. J Supercomput 58:1–31

    Google Scholar 

  47. Zhang Y, Liu R, Wang X, Chen H, Li C (2020) Boosted binary harris hawks optimizer and feature selection. Structure 58(25):26

    Google Scholar 

  48. Elgamal Zenab M, Yasin Norizan BM, Tubishat M, Alswaitti M, Mirjalili S (2020) An improved harris hawks optimization algorithm with simulated annealing for feature selection in the medical field. IEEE Access 8:186638–186652

    Article  Google Scholar 

  49. Salima O, Mohamed AE (2020) Enhanced crow search algorithm for feature selection. Expert Syst Appl 25:113572

    Google Scholar 

  50. Mafarja M, Aljarah I, Heidari AA, Faris H, Fournier-Viger P, Li X, Mirjalili S (2018) Binary dragonfly optimization for feature selection using time-varying transfer functions. Knowl-Based Syst 161:185–204

    Article  Google Scholar 

  51. Zhang H, Wang J, Sun Z, Zurada Jacek M, Pal Nikhil R (2019) Feature selection for neural networks using group lasso regularization. IEEE Trans Knowl Data Eng 32(4):659–673

    Article  Google Scholar 

  52. Mafarja M, Aljarah I, Faris H, Hammouri Abdelaziz I, Ala’M A-Z, Mirjalili S (2019) Binary grasshopper optimisation algorithm approaches for feature selection problems. Expert Syst Appl 117:267–286

    Article  Google Scholar 

  53. Das A, Das S (2017) Feature weighting and selection with a pareto-optimal trade-off between relevancy and redundancy. Pattern Recognit Lett 88:12–19

    Article  Google Scholar 

  54. Whitley D (1994) A genetic algorithm tutorial. Stat Comput 4(2):65–85

    Article  Google Scholar 

  55. Ali Asghar H, Seyedali M, Hossam F, Ibrahim A, Majdi M, Huiling C (2019) Algorithm and applications, Harris hawks optimization. Fut Gener Computer Syst 97:849–872

    Article  Google Scholar 

  56. Eberhart R, Kennedy J (1995) A new optimizer using particle swarm theory. In MHS’95. Proceedings of the Sixth International Symposium on Micro Machine and Human Science, pp. 39–43. Ieee,

  57. Mirjalili S, Gandomi Amir H, Mirjalili Seyedeh Z, Saremi S, Faris H, Mirjalili SM (2017) Salp swarm algorithm: a bio-inspired optimizer for engineering design problems. Adv Eng Softw 114:163–191

    Article  Google Scholar 

  58. Mirjalili S, Lewis A (2016) The whale optimization algorithm. Adv Eng Softwa 95:51–67

    Article  Google Scholar 

  59. Mirjalili S (2015) Moth-flame optimization algorithm: a novel nature-inspired heuristic paradigm. Knowl-Based Syst 89:228–249

    Article  Google Scholar 

  60. Saremi S, Mirjalili S, Lewis A (2017) Grasshopper optimisation algorithm: theory and application. Adv Eng Softw 105:30–47

    Article  Google Scholar 

  61. Dua D, Graff C (2017) UCI machine learning repository,

  62. Algamal ZY, Alhamzawi R, Ali Haithem TM (2018) Gene selection for microarray gene expression classification using bayesian lasso quantile regression. Computers Biol Med 97:145–152

    Article  Google Scholar 

  63. Algamal Zakariya Y, Lee MH, Al-Fakih AM (2016) High-dimensional quantitative structure-activity relationship modeling of influenza neuraminidase a/pr/8/34 (h1n1) inhibitors based on a two-stage adaptive penalized rank regression. J Chemometr 30(2):50–57

    Article  Google Scholar 

  64. Algamal ZY, Lee MH, Al-Fakih AM, Aziz M (2017) High-dimensional qsar classification model for anti-hepatitis c virus activity of thiourea derivatives based on the sparse logistic regression model with a bridge penalty. J Chemometr 31(6):e2889

    Article  Google Scholar 

  65. Algamal ZY, Qasim MK, Ali HTM (2017) A qsar classification model for neuraminidase inhibitors of influenza a viruses (h1n1) based on weighted penalized support vector machine. SAR and QSAR Environ Res 28(5):415–426

    Article  Google Scholar 

  66. Al-Thanoon Niam A, Qasim Omar S, Algamal ZY (2019) A new hybrid firefly algorithm and particle swarm optimization for tuning parameter estimation in penalized support vector machine with application in chemometrics. Chemometr Intell Lab Syst 184:142–152

    Article  Google Scholar 

  67. Al-Dabbagh ZT, Algamal ZY (2019) A robust quantitative structure-activity relationship modelling of influenza neuraminidase a/pr/8/34 (h1n1) inhibitors based on the rank-bridge estimator. SAR and QSAR Environ Res 30(6):417–428

    Article  Google Scholar 

Download references

Acknowledgements

This work was supported by National Natural Science Foundation of China (Grant No. 62150410434) and in part by LIESMARS Special Research Funding.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Ahmed A. Ewees.

Ethics declarations

Conflict of interest

The authors declare that there is no conflict of interest.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Ewees, A.A., Al-qaness, M.A.A., Abualigah, L. et al. Enhanced feature selection technique using slime mould algorithm: a case study on chemical data. Neural Comput & Applic 35, 3307–3324 (2023). https://doi.org/10.1007/s00521-022-07852-8

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s00521-022-07852-8

Keywords