Skip to main content

Advertisement

Log in

A Multi-objective hybrid filter-wrapper evolutionary approach for feature selection

  • Regular Research Paper
  • Published:
Memetic Computing Aims and scope Submit manuscript

Abstract

Feature selection is an important pre-processing data mining task, which can reduce the data dimensionality and improve not only the classification accuracy but also the classifier efficiency. Filters use statistical characteristics of the data as the evaluation measure rather than using a classification algorithm. On the contrary, the wrapper process is computationally expensive because the evaluation of every feature subset requires running the classifier on the datasets and computing the accuracy from the obtained confusion matrix. In order to solve this problem, we propose a hybrid tri-objective evolutionary algorithm that optimizes two filter objectives, namely the number of features and the mutual information, and one wrapper objective corresponding to the accuracy. Once the population is classified into different non-dominated fronts, only feature subsets belonging to the first (best) one are improved using the indicator-based multi-objective local search. Our proposed hybrid algorithm, named Filter-Wrapper-based Nondominated Sorting Genetic Algorithm-II, is compared against several multi-objective and single-objective feature selection algorithms on eighteen benchmark datasets having different dimensionalities. Experimental results show that our proposed algorithm gives competitive and better results with respect to existing algorithms.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4

Similar content being viewed by others

References

  1. Gheyas IA, Smith LS (2010) Feature subset selection in large dimensionality domains. Pattern Recognit 43:5–13

    Article  MATH  Google Scholar 

  2. Kohavi R, John GH (1997) Wrappers for feature subset selection. Artif Intell 97:273–324

    Article  MATH  Google Scholar 

  3. Xue B, Zhang M, Browne WN, Yao X (2016) A Survey on evolutionary computation approaches to feature selection. IEEE Trans Evol Comput 20:606–626

    Article  Google Scholar 

  4. Mukhopadhyay A, Maulik U (2013) An SVM-wrapped multiobjective evolutionary feature selection approach for identifying cancer-microRNA markers. IEEE Trans Nanobiosci 12:275–281

    Article  Google Scholar 

  5. Xue B, Zhang M, Browne WN (2013) Particle swarm optimization for feature selection in classification: a multi-objective approach. IEEE Trans Cybern 43:1656–1671

    Article  Google Scholar 

  6. Hart WE, Krasnogor N, Smith JE (2005) Recent advances in memetic algorithms. Springer, Berlin, p 410. ISBN: 978-3-540-32363-1

  7. Canuto AMP, Nascimento DSC (2012) A genetic-based approach to features selection for ensembles using a hybrid and adaptive fitness function. In: Proceedings of the international joint conference on neural networks (IJCNN), pp 1–8

  8. Liu H, Zhao Z (2015) Manipulating data and dimension reduction methods: feature selection. In: Meyers R (eds) Encyclopedia of complexity and systems science. Springer, Berlin, Heidelberg

  9. Hamdani TM, Won JM, Alimi AM, Karray F (2007) Multi-objective feature selection with NSGA II. In: Beliczynski B, Dzielinski A, Iwanowski M, Ribeiro B (eds) Adaptive and natural computing algorithms. ICANNGA 2007. Lecture Notes in Computer Science, vol 4431. Springer, Berlin, Heidelberg

  10. Huang J, Cai Y, Xu X (2007) A hybrid genetic algorithm for feature selection wrapper based on mutual information. Pattern Recognit Lett 28:1825–1844

    Article  Google Scholar 

  11. Zhu Z, Jia S, Ji Z (2010) Towards a memetic feature selection paradigm [application notes]. IEEE Comput Intell Mag 5:41–53

    Article  Google Scholar 

  12. Yang C-S, Chuang L-Y, Chen Y-J, Yang C-H (2008) Feature selection using memetic algorithms. In: Proceedings of the third international conference on convergence and hybrid information technology (ICCIT ’08), pp 416–423

  13. Butler-Yeoman T, Xue B, Zhang M (2015) Particle swarm optimisation for feature selection: a hybrid filter–wrapper approach. In: Proceedings of the IEEE congress on evolutionary computation (CEC 2015), pp 2428–2435

  14. Deb K, Pratap A, Agarwal S, Meyarivan T (2002) A fast and elitist multiobjective genetic algorithm: NSGA-II. IEEE Trans Evol Comput 6:182–197

    Article  Google Scholar 

  15. Cover T, Hart P (1967) Nearest neighbor pattern classification. IEEE Trans Inf Theory 13:21–27

    Article  MATH  Google Scholar 

  16. Basseur M, Burke EK (2007) Indicator-based multi-objective local search. In: Proceedings of IEEE congress on evolutionary computation, pp 3100–3107

  17. Derrac J, García S, Molina D, Herrera F (2011) A practical tutorial on the use of nonparametric statistical tests as a methodology for comparing evolutionary and swarm intelligence algorithms. Swarm Evol Comput 1:3–18

    Article  Google Scholar 

  18. Frank A, Asuncion A (2010) UCI machine learning repository. https://archive.ics.uci.edu/ml/datasets.html. Accessed Feb 2017

  19. Gütlein M, Frank E, Hall M, Karwath A (2009) Large-scale attribute selection using wrappers. In: Proceedings of the IEEE symposium on computational intelligence and data mining (CIDM ’09), pp 332–339

  20. Abeel T, de Peer YV, Saeys Y (2009) Java-ml: a machine learning library. J Mach Learn Res 10:931–934

    MathSciNet  MATH  Google Scholar 

  21. Shannon C, Weaver W (1948) The mathematical theory of communication. The University of Illinois Press, Urbana, p 144. ISBN: 978-0252725487

  22. Hastie T, Tibshirani R, Friedman J (2009) The elements of statistical learning: data mining, inference and prediction. Springer, New York, p 745. ISBN: 978-0-387-84858-7

  23. Zitzler E, Laumanns M, Thiele L (2001) SPEA2: improving the strength pareto evolutionary algorithm for multiobjective optimization. In: Proceedings of the evolutionary methods for design, optimization and control with applications to industrial problems, pp 95–100

  24. Xue B, Cervante L, Shang L, Browne WN, Zhang M (2013) Multi-objective evolutionary algorithms for filter based feature selection in classification. Int J Artif Intell Tools 22:1350024–31

    Article  Google Scholar 

  25. Bermejo P, Gámez JA, Puerta JM (2011) A GRASP algorithm for fast hybrid (filter–wrapper) feature subset selection in high-dimensional datasets. Pattern Recognit Lett 32:701–711

    Article  Google Scholar 

Download references

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Marwa Hammami.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Appendix A

Appendix A

1.1 A.1 FW-NSGA-II results using C4.5

Table 7 shows that the number of features selected by FW-NSGA-II is smaller than the total number of features when using C4.5 in the wrapper evaluations. It also increases the classification accuracy on almost all low- and medium-dimensional datasets. Compared with WrapperAlgo, FW-NSGA-II outperformed the WrapperAlgo in terms of accuracy on almost all datasets. Comparing FW-NSGA-II with FilterAlgo, FastAlgo, and RapidPSO, FW-NSGA-II outperformed FilterAlgo, FastPSO, and RapidPSO in terms of the classification performance and the size of feature subset in most cases in low- and medium-dimensional datasets. FW-NSGA-II tends to produce larger feature subsets than WrapperAlgo, GRASP and Random Forest, but smaller subsets than FastPSO, RapidPSO, and FilterAlgo. FW-NSGA-II outperformed the GRASP and the RF algorithm in terms of accuracy in almost all low- and medium-dimensional datasets.

For high-dimensional datasets, FW-NSGA-II outperformed FilterAlgo, FastAlgo, RapidPSO and GRASP in terms of the classification performance and the number of features. However, the Random Forest algorithm achieved better performance than FW-NSGA-II and all the other competitive algorithms in terms of the number of features and the classification accuracy in almost all datasets. The number of features selected by FW-NSGA-II in high-dimensional datasets is higher than the number of features selected by FW-NSGA-II in low- and medium-dimensional datasets. In fact, the number of removed features in FW-NSGA-II for the Ovarian dataset is around 20%, which is different from the number of removed features in low- and medium-dimensional datasets which is around 50% of the original feature size.

The Decision Tree classifier improved the performance of FW-NSGA-II in terms of the classification accuracy and the number of features, compared with the K-NN classifier. The results show that FW-NSGA-II achieved better results than the other algorithms in almost all low- and medium-dimensional datasets, but this is not the case with high-dimensional datasets. In fact, the Random Forest algorithm outperformed the FW-NSGA-II due to its ability to handle large datasets with higher dimensionality compared with the Decision Tree classifier. The source code and the experimental data could be obtained upon request from the corresponding author.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Hammami, M., Bechikh, S., Hung, CC. et al. A Multi-objective hybrid filter-wrapper evolutionary approach for feature selection. Memetic Comp. 11, 193–208 (2019). https://doi.org/10.1007/s12293-018-0269-2

Download citation

  • Received:

  • Accepted:

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s12293-018-0269-2

Keywords

Navigation