Abstract
Feature selection is an essential stage in many data mining and machine learning and applications that find the proper subset of features from a set of irrelevant, redundant, noisy and high dimensional data. This dimensional reduction is a vital task to increase classification accuracy and thus reduce the processing time. An optimization algorithm can be applied to tackle the feature selection problem. In this paper, a \(\beta\)-hill climbing optimizer is applied to solve the feature selection problem. \(\beta\)-hill climbing is recently introduced as a local-search based algorithm that can obtain pleasing solutions for different optimization problems. In order to tailor \(\beta\)-hill climbing for feature selection, it has to be adapted to work in a binary context. The S-shaped transfer function is used to transform the data into the binary representation. A set of 22 de facto benchmark real-world datasets are used to evaluate the proposed algorithm. The effect of the \(\beta\)-hill climbing parameters on the convergence rate is studied in terms of accuracy, the number of features, fitness values, and computational time. Furthermore, the proposed method is compared against three local search methods and ten metaheuristics methods. The obtained results show that the proposed binary \(\beta\)-hill climbing optimizer outperforms other comparative local search methods in terms of classification accuracy on 16 out of 22 datasets. Furthermore, it overcomes other comparative metaheuristics approaches in terms of classification accuracy in 7 out of 22 datasets. The obtained results prove the efficiency of the proposed binary \(\beta\)-hill climbing optimizer.


Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.References
Abed-alguni B, Klaib A (2018) Hybrid whale optimization and \(\beta\)-hill climbing algorithm. Int J CompuT Sci Math, pp 1–13
Abed-alguni BH, Alkhateeb F (2018) Intelligent hybrid cuckoo search and \(\beta\)-hill climbing algorithm. J King Saud Univ Comput Inf Sci 32(2):159–173
Abualigah LM, Khader AT, Al-Betar MA (2017a) \(\beta\)-hill climbing technique for the text document clustering. New Trends in Information Technology NTIT2017 Conference, Amman, Jordan, IEEE, pp 1–6
Abualigah LM, Khadery AT, Al-Betar MA, Alyasseri ZAA, Alomari OA, Hanandehk ES (2017b) Feature selection with \(\beta\)-hill climbing search for text clustering application. Second Palestinian International Conference on Information and Communication Technology (PICICT 2017), Gaza, Palestine, IEEE, pp 22–27
Al-Abdallah RZ, Jaradat AS, Doush IA, Jaradat YA (2017) Abinary classifier based on firefly algorithm. Jordan J Comput Inf Technol (JJCIT) 3(3)
Al-Betar MA (2017) \(\beta\)-hill climbing: an exploratory local search. Neural Comput Appl 28(1):153–168. https://doi.org/10.1007/s00521-016-2328-2
Al-Betar MA, Awadallah MA, Bolaji AL, Alijla BO (2017) \(\beta\)-hill climbing algorithm for sudoku game. Second Palestinian International Conference on Information and Communication Technology (PICICT 2017), Gaza, Palestine, IEEE, pp 84–88
Al-Betar MA, Awadallah MA, Abu Doush I, Alsukhni E, ALkhraisat H (2018) A non-convex economic dispatch problem with valve loading effect using a new modified \(\beta\)-hill climbing local search algorithm. Arab J Sci Eng. https://doi.org/10.1007/s13369-018-3098-1
Aljarah I, Mafarja M, Heidari AA, Faris H, Zhang Y, Mirjalili S (2018) Asynchronous accelerating multi-leader salp chains for feature selection. Appl Soft Comput 71:964–979
Alomari OA, Khader AT, Al-Betar MA, Alyasseri ZAA (2018a) A hybrid filter-wrapper gene selection method for cancer classification. 2018 2nd International Conference on BioSignal Analysis. Processing and Systems (ICBAPS), IEEE, pp 113–118
Alomari OA, Khader AT, Al-Betar MA, Awadallah MA (2018b) A novel gene selection method using modified mrmr and hybrid bat-inspired algorithm with \(\beta\)-hill climbing. Appl Intell 48(11):4429–4447
Alsaafin A, Elnagar A (2017) A minimal subset of features using feature selection for handwritten digit recognition. J Intell Learn Syst Appl 9(4):55–68
Alsukni E, Arabeyyat OS, Awadallah MA, Alsamarraie L, Abu-Doush I, Al-Betar MA (2017) Multiple-reservoir scheduling using \(\beta\)-hill climbing algorithm. J Intell Syst 28(4):559–570
Alyasseri ZAA, Khader AT, Al-Betar MA (2017a) Optimal eeg signals denoising using hybrid \(\beta\)-hill climbing algorithm and wavelet transform. ICISPC ’17. Penang, Malaysia, ACM, pp 5–11
Alyasseri ZAA, Khader AT, Al-Betar MA (2017b) Optimal electroencephalogram signals denoising using hybrid \(\beta\)-hill climbing algorithm and wavelet transform. In: Proceedings of the International Conference on Imaging, Signal Processing and Communication, ACM, pp 106–112
Alyasseri ZAA, Khader AT, Al-Betar MA, Awadallah MA (2018) Hybridizing \(\beta\)-hill climbing with wavelet transform for denoising ecg signals. Inf Sci 429:229–246
Alzaidi AA, Ahmad M, Doja MN, Al Solami E, Beg MS (2018) A new 1d chaotic map and \(beta\)-hill climbing for generating substitution-boxes. IEEE Access 6:55405–55418
Bermejo P, Gámez JA, Puerta JM (2011) A grasp algorithm for fast hybrid (filter-wrapper) feature subset selection in high-dimensional datasets. Pattern Recognit Lett 32(5):701–711
Bermejo P, Gámez JA, Puerta JM (2014) Speeding up incremental wrapper feature subset selection with naive bayes classifier. Knowl-Based Syst 55(Supplement C):14–147, https://doi.org/10.1016/j.knosys.2013.10.016, http://www.sciencedirect.com/science/article/pii/S0950705113003274
Blum C, Roli A (2003) Metaheuristics in combinatorial optimization: overview and conceptual comparison. ACM Comput Surv (CSUR) 35(3):268–308
Bolón-Canedo V, Alonso-Betanzos A (2019) Ensembles for feature selection: a review and future trends. Inf Fusion 52:1–12
Boughaci D, Alkhawaldeh AAs (2018) Three local search-based methods for feature selection in credit scoring. Vietnam J Comput Sci 5(2):107–121
Chen Y, Garcia EK, Gupta MR, Rahimi A, Cazzanti L (2009) Similarity-based classification: concepts and algorithms. J Mach Learn Res 10(Mar):747–776
Corana A, Marchesi M, Martini C, Ridella S (1987) Minimizing multimodal functions of continuous variables with the “simulated annealing” algorithm corrigenda for this article is available here. ACM Trans Math Softw (TOMS) 13(3):262–280
Cover TM, Thomas JA (2012) Elements of information theory. Wiley, Hoboken
Delen D, Walker G, Kadam A (2005) Predicting breast cancer survivability: a comparison of three data mining methods. Artif Intell Med 34(2):113–127
Deng X, Li Y, Weng J, Zhang J (2019) Feature selection for text classification: a review. Multimed Tools Appl 78(3):3797–3816
Doush IA, Sahar AB (2017) Currency recognition using a smartphone: comparison between color sift and gray scale sift algorithms. J King Saud Univ Comput Inf Sci 29(4):484–492
Dubey SR, Singh SK, Singh RK (2015) Local wavelet pattern: a new feature descriptor for image retrieval in medical ct databases. IEEE Trans Image Process 24(12):5892–5903
Duda RO, Hart PE, Stork DG (2012) Pattern classification. Wiley, Hoboken
Emary E, Zawbaa HM, Hassanien AE (2016) Binary ant lion approaches for feature selection. Neurocomputing 213:54–65
Forman G (2003) An extensive empirical study of feature selection metrics for text classification. J Mach Learn Res 3(Mar):1289–1305
García S, Fernández A, Luengo J, Herrera F (2010) Advanced nonparametric tests for multiple comparisons in the design of experiments in computational intelligence and data mining: Experimental analysis of power. Inf Sci 180(10):2044–2064
Ghareb AS, Bakar AA, Hamdan AR (2016) Hybrid feature selection based on enhanced genetic algorithm for text categorization. Expert Syst Appl 49:31–47
Gheyas IA, Smith LS (2010) Feature subset selection in large dimensionality domains. Pattern Recognit 43(1):5–13. https://doi.org/10.1016/j.patcog.2009.06.009, http://www.sciencedirect.com/science/article/pii/S0031320309002520
Goltsev A, Gritsenko V (2012) Investigation of efficient features for image recognition by neural networks. Neural Netw 28:15–23
Gu S, Cheng R, Jin Y (2018) Feature selection for high-dimensional classification using a competitive swarm optimizer. Soft Comput 22(3):811–822
Hall MA, Smith LA (1999) Feature selection for machine learning: comparing a correlation-based filter approach to the wrapper. FLAIRS Conf 1999:235–239
Hu Q, Yu D, Xie Z (2006) Information-preserving hybrid data reduction based on fuzzy-rough techniques. Pattern Recognit Lett 27(5):414–423
Jing LP, Huang HK, Shi HB (2002) Improved feature selection approach tfidf in text mining. In: Proceedings. International Conference on Machine Learning and Cybernetics, IEEE, vol 2, pp 944–946
Kabir MM, Shahjahan M, Murase K (2012) A new hybrid ant colony optimization algorithm for feature selection. Expert Syst Appl 39(3):3747–3763
Kashef S, Nezamabadi-pour H (2015) An advanced aco algorithm for feature subset selection. Neurocomputing 147:271–279
Kennedy J, Eberhart RC (1997) A discrete binary version of the particle swarm algorithm. In: 1997 IEEE International conference on systems, man, and cybernetics. Computational cybernetics and simulation, IEEE, vol 5, pp 4104–4108
Lai C, Reinders MJ, Wessels L (2006) Random subspace method for multivariate feature selection. Pattern Recognit Lett 27(10):1067–1076
Lee C, Lee GG (2006) Information gain and divergence-based feature selection for machine learning-based text categorization. Inf Process Manag 42(1):155–165
Li P, Shrivastava A, Moore JL, König AC (2011) Hashing algorithms for large-scale learning. In: Advances in neural information processing systems, pp 2672–2680
Li Y, Li T, Liu H (2017) Recent advances in feature selection and its applications. Knowl Inf Syst 53(3):551–577
Liao Y, Vemuri VR (2002) Use of k-nearest neighbor classifier for intrusion detection. Comput Secur 21(5):439–448
Liu H, Setiono R (1995) Chi2: Feature selection and discretization of numeric attributes, pp 388–391
Ma B, Xia Y (2017) A tribe competition-based genetic algorithm for feature selection in pattern classification. Appl Soft Comput 58(Supplement C):328–338, https://doi.org/10.1016/j.asoc.2017.04.042, http://www.sciencedirect.com/science/article/pii/S1568494617302247
Mafarja M, Mirjalili S (2018) Whale optimization approaches for wrapper feature selection. Appl Soft Comput 62:441–453
Mafarja M, Aljarah I, Heidari AA, Faris H, Fournier-Viger P, Li X, Mirjalili S (2018a) Binary dragonfly optimization for feature selection using time-varying transfer functions. Knowl-Based Syst 161:185–204
Mafarja M, Aljarah I, Heidari AA, Hammouri AI, Faris H, Ala’M AZ, Mirjalili S (2018b) Evolutionary population dynamics and grasshopper optimization approaches for feature selection problems. Knowl-Based Syst 145:25–45
Mafarja M, Aljarah I, Faris H, Hammouri AI, Ala’M AZ, Mirjalili S (2019) Binary grasshopper optimisation algorithm approaches for feature selection problems. Expert Syst Appl 117:267–286
Marinaki M, Marinakis Y (2015) A hybridization of clonal selection algorithm with iterated local search and variable neighborhood search for the feature selection problem. Memetic Comput 7(3):181–201
Mashrgy MA, Bdiri T, Bouguila N (2014) Robust simultaneous positive data clustering and unsupervised feature selection using generalized inverted dirichlet mixture models. Knowl-Based Syst 59(Supplement C):182–195, https://doi.org/10.1016/j.knosys.2014.01.007, http://www.sciencedirect.com/science/article/pii/S0950705114000185
Mlakar U, Fister I, Brest J, Potočnik B (2017) Multi-objective differential evolution for feature selection in facial expression recognition systems. Expert Syst Appl 89:129–137
Moayedikia A, Ong KL, Boo YL, Yeoh WG, Jensen R (2017) Feature selection for high dimensional imbalanced class data using harmony search. Eng Appl Artif Intell 57:38–49
Park CH, Kim SB (2015) Sequential random k-nearest neighbor feature selection for high-dimensional data. Expert Syst Appl 42(5):2336–2342. https://doi.org/10.1016/j.eswa.2014.10.044, http://www.sciencedirect.com/science/article/pii/S095741741400668X
Rashedi E, Nezamabadi-Pour H, Saryazdi S (2013) A simultaneous feature adaptation and feature selection method for content-based image retrieval systems. Knowl-Based Syst 39:85–94
Ravisankar P, Ravi V, Rao GR, Bose I (2011) Detection of financial statement fraud and feature selection using data mining techniques. Decision Support Syst 50(2):491–500
Robnik-Šikonja M, Kononenko I (2003) Theoretical and empirical analysis of relieff and rrelieff. Mach Learn 53(1):23–69
Saeys Y, Inza I, Larrañaga P (2007) A review of feature selection techniques in bioinformatics. Bioinformatics 23(19):2507–2517
Sawalha R, Doush IA (2012) Face recognition using harmony search-based selected features. Int J Hybrid Inf Technol 5(2):1–16
Sayed GI, Hassanien AE, Azar AT (2019) Feature selection via a novel chaotic crow search algorithm. Neural Comput Appl 31(1):171–188
Shang C, Li M, Feng S, Jiang Q, Fan J (2013) Feature selection via maximizing global information gain for text classification. Knowl-Based Syst 54(Supplement C):298 – 309, https://doi.org/10.1016/j.knosys.2013.09.019, http://www.sciencedirect.com/science/article/pii/S0950705113003067
Shao C, Paynabar K, Kim TH, Jin JJ, Hu SJ, Spicer JP, Wang H, Abell JA (2013) Feature selection for manufacturing process monitoring using cross-validation. J Manuf Syst 32(4):550–555
Sindhu SSS, Geetha S, Kannan A (2012) Decision tree based light weight intrusion detection using a wrapper approach. Expert Syst Appl 39(1):129–141. https://doi.org/10.1016/j.eswa.2011.06.013, http://www.sciencedirect.com/science/article/pii/S0957417411009080
Talbi EG (2009) Metaheuristics: from design to implementation, vol 74. Wiley, Hoboken
Taradeh M, Mafarja M, Heidari AA, Faris H, Aljarah I, Mirjalili S, Fujita H (2019) An evolutionary gravitational search-based feature selection. Inf Sci 497:219–239
Urbanowicz RJ, Olson RS, Schmitt P, Meeker M, Moore JH (2018) Benchmarking relief-based feature selection methods for bioinformatics data mining. J Biomed Inform 85:168–188
Weinberger KQ, Saul LK (2009) Distance metric learning for large margin nearest neighbor classification. J Mach Learn Res 10(Feb):207–244
Wieland M, Pittore M (2014) Performance evaluation of machine learning algorithms for urban pattern recognition from multi-spectral satellite images. Remote Sens 6(4):2912–2939
Wolpert DH, Macready WG et al (1997) No free lunch theorems for optimization. IEEE Trans Evol Comput 1(1):67–82
Yu L, Liu H (2003) Feature selection for high-dimensional data: a fast correlation-based filter solution. In: Proceedings of the 20th international conference on machine learning (ICML-03), pp 856–863
Zhang H, Sun G (2002) Feature selection using tabu search method. Pattern Recognit 35(3):701–711
Zhang L, Mistry K, Lim CP, Neoh SC (2018) Feature selection using firefly optimization for classification and regression models. Decision Support Syst 106:64–85
Zhang Y, Wang S, Phillips P, Ji G (2014) Binary pso with mutation operator for feature selection using decision tree applied to spam detection. Knowl-Based Syst 64:22–31
Zhao Z, Liu H (2007) Spectral feature selection for supervised and unsupervised learning. In: Proceedings of the 24th international conference on Machine learning, ACM, pp 1151–1157
Zhong N, Dong J, Ohsuga S (2001) Using rough sets with heuristics for feature selection. J Intell Inf Syst 16(3):199–214
Author information
Authors and Affiliations
Corresponding author
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Al-Betar, M.A., Hammouri, A.I., Awadallah, M.A. et al. Binary \(\beta\)-hill climbing optimizer with S-shape transfer function for feature selection. J Ambient Intell Human Comput 12, 7637–7665 (2021). https://doi.org/10.1007/s12652-020-02484-z
Received:
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s12652-020-02484-z