Skip to main content

Using Clustering for Supervised Feature Selection to Detect Relevant Features

  • Conference paper
  • First Online:
Machine Learning, Optimization, and Data Science (LOD 2019)

Part of the book series: Lecture Notes in Computer Science ((LNISA,volume 11943))

Abstract

In many applications in machine learning, large quantities of features and information are available, but these can be of low quality. A novel filter method for feature selection for classification termed COLD is presented that uses class-wise clustering to reduce the dimensionality of the data. The idea behind this approach is that if a relevant feature would be removed from the set of features, the separation of clusters belonging to different classes will deteriorate. Four artificial examples and two real-world data sets are presented on which COLD is compared with several popular filter methods. For the artificial examples, only COLD is capable to consistently rank the features according to their contribution to the separation of the classes. For the real-world Dermatology and Arrhythmia dataset, COLD demonstrates the ability to remove a large number of features and improve the classification accuracy or, at a minimum, not degrade the performance considerably.

This is a preview of subscription content, log in via an institution to check access.

Access this chapter

Chapter
USD 29.95
Price excludes VAT (USA)
  • Available as PDF
  • Read on any device
  • Instant download
  • Own it forever
eBook
USD 39.99
Price excludes VAT (USA)
  • Available as EPUB and PDF
  • Read on any device
  • Instant download
  • Own it forever
Softcover Book
USD 54.99
Price excludes VAT (USA)
  • Compact, lightweight edition
  • Dispatched in 3 to 5 business days
  • Free shipping worldwide - see info

Tax calculation will be finalised at checkout

Purchases are for personal use only

Institutional subscriptions

References

  1. Bishop, C.M.: Pattern Recognition and Machine Learning. Springer ScienceBusiness Media, New York (2006)

    MATH  Google Scholar 

  2. Caruana, R., Freitag, D.: Greedy attribute selection. In: Cohen, W., Hirsh, H. (eds.) Proceedings of the 11th International Conference on Machine Learning (ICML 1994), pp. 28–36. Morgan Kaufmann, New Brunswick (1994)

    Google Scholar 

  3. Chan, T.F.: Rank revealing QR factorizations. Linear Algebra Appl. 88–89, 67–82 (1987)

    MathSciNet  MATH  Google Scholar 

  4. Chormunge, S., Jena, S.: Correlation based feature selection with clustering for highdimensional data. J. Electr. Syst. Inf. Technol. 5, 542–549 (2018)

    Google Scholar 

  5. Cover, T.M.: The best two independent measurements are not the two best. IEEE Trans. Syst. Man Cybern. 4(1), 116–117 (1974)

    Article  Google Scholar 

  6. Dessì, N., Pes, B.: Similarity of feature selection methods. An empirical study across data intensive classification tasks. Expert Syst. Appl. 42(10), 4632–4642 (2015)

    Article  Google Scholar 

  7. Duda, R., Hart, P., Stork, D.: Pattern Classification. John Wiley and Sons, New York (2012)

    MATH  Google Scholar 

  8. Elashoff, J.E., Elashoff, R.M., Goldman, G.E.: On the choice of variables in classification problems with dichotomous variables. Biometrika 54(3), 668–670 (1967)

    Article  MathSciNet  Google Scholar 

  9. Ruffo, G.: Matlab Toolbox: Feature selection library. https://se.mathworks.com/matlabcentral/fileexchange/56937-feature-selection-library. Accessed 1 Dec 2018

  10. Guyon, I., Elisseeff, A.: An introduction to variable and feature selection. J. Mach. Learn. Res. 3, 1157–1182 (2003)

    MATH  Google Scholar 

  11. Hastie, T., Tibshirani, R., Friedman, J.: Data Mining, Inference, and Prediction. Springer Series in Statistics. Springer, New York (2009). https://doi.org/10.1007/978-0-387-84858-7

    Book  MATH  Google Scholar 

  12. He X., Cai, D., Niyogi, P.: Laplacian score for feature selection. In: Proceedings of the 18th International Conference on Neural Information Processing Systems (NIPS 2005), pp. 507–514. MIT Press, Cambridge (2005)

    Google Scholar 

  13. Kittler, J., Mardia, K.V.: Statistical pattern recognition in image analysis. J. Appl. Stat. 21(1–2), 61–75 (1994)

    Article  Google Scholar 

  14. Kononenko, I.: Estimating attributes: analysis and extensions of RELIEF. In: Bergadano, F., De Raedt, L. (eds.) ECML 1994. LNCS, vol. 784, pp. 171–182. Springer, Heidelberg (1994). https://doi.org/10.1007/3-540-57868-4_57

    Chapter  Google Scholar 

  15. Kononenko, I., Simec, E., Robnik-Sikonja, M.: Overcoming the myopia of inductive learning algorithms with RELIEFF. Appl. Intell. 7, 39–55 (1997)

    Article  Google Scholar 

  16. Lichman, M.: UCI Machine Learning Repository. https://archive.ics.uci.edu/ml/index.php. Accessed 20 June 2019

  17. Lohrmann, C., Luukka, P., Jablonska-Sabuka, M., Kauranne, T.: Supervised feature selection with a combination of fuzzy similarity measures and fuzzy entropy measures. Expert Syst. Appl. 110, 216–236 (2018)

    Article  Google Scholar 

  18. Luukka, P.: Feature selection using fuzzy entropy measures with similarity classifier. Expert Syst. Appl. 38, 4600–4607 (2011)

    Article  Google Scholar 

  19. Mitra, P., Murthy, C.A., Pal, S.K.: Unsupervised feature selection using feature similarity. IEEE Trans. Pattern Anal. Mach. Intell. 24(3), 301–312 (2002)

    Article  Google Scholar 

  20. Robnik-Šikonja, M., Kononenko, I.: Theoretical and empirical analysis of ReliefF and RReliefF. Appl. Intell. 53(1–2), 23–69 (2003)

    MATH  Google Scholar 

  21. Rousseeuw, P.J.: A graphical aid to the interpretation and validation of cluster analysis. J. Comput. Appl. Math. 20, 53–65 (1987)

    Article  Google Scholar 

  22. Sahu, B., Dehuri, S., Jagadev, A.K.: Feature selection model based on clustering and ranking in pipeline for microarray data. Inf. Med. Unlocked 9, 107–122 (2017)

    Article  Google Scholar 

  23. Sammut, C., Webb, G.I.: Encyclopedia of Machine Learning and Data Mining, 2017th edn. Springer Science+Business Media, New York (2017)

    Book  Google Scholar 

  24. Sotoca, J.M., Pla, F.: Supervised feature selection by clustering using conditional mutual information-based distances. Pattern Recogn. 43, 2068–2081 (2010)

    Article  Google Scholar 

  25. Toussaint, G.T.: Note on optimal selection of independent binary-valued features for pattern recognition. IEEE Trans. Inf. Theory 17(5), 618 (1971)

    Google Scholar 

  26. Warton, D.I.: Penalized normal likelihood and ridge regularization of correlation and covariance matrices. J. Am. Stat. Assoc. 103(481), 340–349 (2008)

    Article  MathSciNet  Google Scholar 

Download references

Acknowledgements

This research would like to acknowledge the funding received from the Finnish Strategic Research Council, grant number 313396/MFG40 Manufacturing 4.0.

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Christoph Lohrmann .

Editor information

Editors and Affiliations

Rights and permissions

Reprints and permissions

Copyright information

© 2019 Springer Nature Switzerland AG

About this paper

Check for updates. Verify currency and authenticity via CrossMark

Cite this paper

Lohrmann, C., Luukka, P. (2019). Using Clustering for Supervised Feature Selection to Detect Relevant Features. In: Nicosia, G., Pardalos, P., Umeton, R., Giuffrida, G., Sciacca, V. (eds) Machine Learning, Optimization, and Data Science. LOD 2019. Lecture Notes in Computer Science(), vol 11943. Springer, Cham. https://doi.org/10.1007/978-3-030-37599-7_23

Download citation

  • DOI: https://doi.org/10.1007/978-3-030-37599-7_23

  • Published:

  • Publisher Name: Springer, Cham

  • Print ISBN: 978-3-030-37598-0

  • Online ISBN: 978-3-030-37599-7

  • eBook Packages: Computer ScienceComputer Science (R0)

Publish with us

Policies and ethics