Skip to main content

Feature Ranking for Feature Sorting and Feature Selection, and Feature Sorting: FR4(FSoFS)\(\wedge \)FSo

  • Conference paper
  • First Online:
17th International Conference on Soft Computing Models in Industrial and Environmental Applications (SOCO 2022) (SOCO 2022)

Abstract

This paper introduces the application of feature ranking with a twofold purpose: first it achieves a feature sorting which becomes into a feature selection procedure given that a threshold is defined to only retain features with a positive influence with the class label (feature sorting and feature selection: FSoFS), second the feature subset is optimised using feature ranking to get a promising order of attributes (FSo). The supporting paper introduced a single stage: Feature ranking for feature sorting and feature selection, FR4(FS)\(^2\) with the shortcoming that for some high-dimensional classification problems the pre-processed data set did not obtain a more accurate classification model than the raw classifier. The results mean that it deserves to consider feature ranking in a couple of sequential stages with different goals: i) feature sorting, accompanied by feature selection, and ii) also alone

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 129.00
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 169.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Ayele, W.Y.: A toolbox for idea generation and evaluation: machine learning, data-driven, and contest-driven approaches to support idea generation. arXiv preprint arXiv:2205.09840 (2022)

  2. Cardoso, R.P., Hart, E., Kurka, D.B., Pitt, J.: WILDA: Wide Learning of Diverse Architectures for Classification of Large Datasets. In: Castillo, P.A., Jiménez Laredo, J.L. (eds.) EvoApplications 2021. LNCS, vol. 12694, pp. 649–664. Springer, Cham (2021). https://doi.org/10.1007/978-3-030-72699-7_41

    Chapter  Google Scholar 

  3. Huber, S., Wiemer, H., Schneider, D., Ihlenfeldt, S.: Dmme: data mining methodology for engineering applications-a holistic extension to the crisp-dm model. Procedia CIRP 79, 403–408 (2019)

    Article  Google Scholar 

  4. Janošcová, R.: Mining big data in weka. In: 11th IWKM, Bratislava, Slovakia (2016)

    Google Scholar 

  5. Jiang, Z., Zhang, Y., Wang, J.: A multi-surrogate-assisted dual-layer ensemble feature selection algorithm. Appl. Soft Comput. 110, 107625 (2021)

    Article  Google Scholar 

  6. Kuhn, M., Johnson, K.: Data Pre-processing. In: Applied Predictive Modeling, pp. 27–59. Springer, New York (2013). https://doi.org/10.1007/978-1-4614-6849-3_3

    Chapter  MATH  Google Scholar 

  7. Lapucci, M., Pucci, D.: Mixed-integer quadratic programming reformulations of multi-task learning models. Math. Eng. 5(1), 1–16 (2023)

    Article  MathSciNet  Google Scholar 

  8. Kee, G., et al.: Smart robust feature selection (soft) for imbalanced and heterogeneous data. Knowl.-Based Syst. 236, 107197 (2022)

    Google Scholar 

  9. Li, J., Yaoyang, W., Fong, S., Tallón-Ballesteros, A.J., Yang, X.-S., Mohammed, S., Feng, W.: A binary pso-based ensemble under-sampling model for rebalancing imbalanced training data. J. Supercomput. 78(5), 7428–7463 (2022)

    Article  Google Scholar 

  10. Li, J., et al.: Feature selection: a data perspective. ACM Comput. Surv. (CSUR) 50(6), 1–45 (2017)

    Article  Google Scholar 

  11. Liu, H., Motoda, H. Computational methods of feature selection. CRC Press (2007)

    Google Scholar 

  12. Merchán, A.F., Márquez-Rodríguez, A., Santana-Morales, P., Tallón-Ballesteros, A.J.: Feature ranking merging: Frmgg. application in high dimensionality binary classification problems. In: Mathur, G., Bundele, M., Tripathi, A., Paprzycki M. (eds.) ICAIAA 2022. AIS. Springer (2023)

    Google Scholar 

  13. Milne. L.: Feature selection using neural networks with contribution measures. In: Yao, X. (ed.) AI ’95 CONFERENCE-, pp. 571–571. World Scientific (1995)

    Google Scholar 

  14. Nguyen, H.B., Xue, B., Andreae, P.: Pso with surrogate models for feature selection: static and dynamic clustering-based methods. Memetic Comput. 10(3), 291–300 (2018)

    Article  Google Scholar 

  15. Nicolas, P.R.: Scala for machine learning. Packt Publishing Ltd. (2015)

    Google Scholar 

  16. Pramanik, P.K.D., Pal, S., Mukhopadhyay, M., Singh, S.P.: Big data classification: techniques and tools. In: Applications of Big Data in Healthcare, pp. 1–43. Elsevier (2021)

    Google Scholar 

  17. Quinlan, J.R.: Induction of decision trees. pp. 81–106 (1986)

    Google Scholar 

  18. Rodríguez-Molina, A., et al.: Exploiting evolutionary computation techniques for service industries. In: Evolutionary Computation with Intelligent Systems, pp. 153–179. CRC Press

    Google Scholar 

  19. Santana-Morales, P., Merchán, A.F., Márquez-Rodríguez, A., Tallón-Ballesteros, A.J.: Feature ranking for feature sorting and feature selection: Fr4(fs)\({}^{\text{2}}\). In: de Vicente, J.M.F., Álvarez Sánchez, J.R., de la Paz López, F., Adeli, H. (eds.) Bio-inspired Systems and Applications: from Robotics to Ambient Intelligence - 9th International Work-Conference on the Interplay Between Natural and Artificial Computation, IWINAC 2022. LNCS, Part II, vol. 13259, pp. 545–550. Springer, Cham (2022)

    Google Scholar 

  20. Slowik, A., Bottou, L.: On distributionally robust optimization and data rebalancing. In: Camps-Valls, G., Ruiz, F.J.R., Valera, I. (eds.) International Conference on Artificial Intelligence and Statistics, AISTATS 2022, 28–30 March 2022, Virtual Event. Proceedings of Machine Learning Research, vol. 151, pp. 1283–1297. PMLR (2022)

    Google Scholar 

  21. Tallón-Ballesteros, A.J., Riquelme, J.C., Ruiz, R.: Semi-wrapper feature subset selector for feed-forward neural networks: applications to binary and multi-class classification problems. Neurocomputing 353, 28–44 (2019)

    Article  Google Scholar 

  22. Tallón-Ballesteros, A.J., Cavique, L., Fong, S.: Addressing low dimensionality feature subset selection: ReliefF(-k) or extended correlation-based feature selection(eCFS)? In: Martínez Álvarez, F., Troncoso Lora, A., Sáez Muñoz, J.A., Quintián, H., Corchado, E. (eds.) SOCO 2019. AISC, vol. 950, pp. 251–260. Springer, Cham (2020). https://doi.org/10.1007/978-3-030-20055-8_24

    Chapter  Google Scholar 

  23. Tallón-Ballesteros, A.J., Correia, L., Cho, S.-B.: Stochastic and non-stochastic feature selection. In: Yin, H., Gao, Y., Chen, S., Wen, Y., Cai, G., Gu, T., Du, J., Tallón-Ballesteros, A.J., Zhang, M. (eds.) IDEAL 2017. LNCS, vol. 10585, pp. 592–598. Springer, Cham (2017). https://doi.org/10.1007/978-3-319-68935-7_64

    Chapter  Google Scholar 

  24. Tallón-Ballesteros, A.J., Correia, L., Leal-Díaz, R.: Attribute Subset Selection for Image Recognition. Random Forest Under Assessment. In: Sanjurjo González, H., Pastor López, I., García Bringas, P., Quintián, H., Corchado, E. (eds.) SOCO 2021. AISC, vol. 1401, pp. 821–827. Springer, Cham (2022). https://doi.org/10.1007/978-3-030-87869-6_78

    Chapter  Google Scholar 

  25. Tallón-Ballesteros, A.J., Fong, S., Leal-Díaz, R.: Does the order of attributes play an important role in classification? In: Pérez García, H., Sánchez González, L., Castejón Limas, M., Quintián Pardo, H., Corchado Rodríguez, E. (eds.) HAIS 2019. LNCS (LNAI), vol. 11734, pp. 370–380. Springer, Cham (2019). https://doi.org/10.1007/978-3-030-29859-3_32

    Chapter  Google Scholar 

  26. Tallón-Ballesteros, A.J., Riquelme, J.C.: Data mining methods applied to a digital forensics task for supervised machine learning. In: Muda, A.K., Choo, Y.-H., Abraham, A., N. Srihari, S. (eds.) Computational Intelligence in Digital Forensics: Forensic Investigation and Applications. SCI, vol. 555, pp. 413–428. Springer, Cham (2014). https://doi.org/10.1007/978-3-319-05885-6_17

    Chapter  Google Scholar 

  27. Tallón-Ballesteros, A.J., Riquelme, J.C., Ruiz, R.: Filter-based feature selection in the context of evolutionary neural networks in supervised machine learning. Pattern Anal. Appl. 23(1), 467–491 (2019). https://doi.org/10.1007/s10044-019-00798-z

    Article  MathSciNet  Google Scholar 

  28. Wahab, S.A.: Moving towards society 5.0: A bibliometric and visualization analysis. In SOCIETY 5.0: First International Conference, Society 5.0 2021, Virtual Event, p. 93. Springer Nature (2022)

    Google Scholar 

  29. Zhan, Z.-H., Shi, L., Tan, K.C., Zhang, J.: A survey on evolutionary computation for complex continuous optimization. Artif. Intell. Rev. 55(1), 59–110 (2021). https://doi.org/10.1007/s10462-021-10042-y

    Article  Google Scholar 

Download references

Acknowledgements

This work has been partially subsidised by the project US-1263341 (Junta de Andalucía) and FEDER funds.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Antonio J. Tallón-Ballesteros .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2023 The Author(s), under exclusive license to Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Tallón-Ballesteros, A.J., Márquez-Rodríguez, A., Wu, Y., Santana-Morales, P., Fong, S. (2023). Feature Ranking for Feature Sorting and Feature Selection, and Feature Sorting: FR4(FSoFS)\(\wedge \)FSo. In: García Bringas, P., et al. 17th International Conference on Soft Computing Models in Industrial and Environmental Applications (SOCO 2022). SOCO 2022. Lecture Notes in Networks and Systems, vol 531. Springer, Cham. https://doi.org/10.1007/978-3-031-18050-7_56

Download citation

Publish with us

Policies and ethics