Skip to main content
Log in

\(\hbox {U}^2\hbox {F}^2\hbox {S}^2\): Uncovering Feature-level Similarities for Unsupervised Feature Selection

  • Published:
Neural Processing Letters Aims and scope Submit manuscript

Abstract

Unsupervised feature selection is a critical technique in processing high dimensional data containing redundant and noisy features. Based on sample-level similarities, conventional algorithms select features that can preserve the local structure of data points. However, the similarities among all dimensions of features, which play important roles in feature selection, are neglected. In this paper, we propose a novel method dubbed \(\hbox {U}^2\hbox {F}^2\hbox {S}^2\) by uncovering these pivotal similarities for unsupervised feature selection. A feature-level similarity uncovering loss function is first presented to preserve the local structure of data points at the feature level. Specially, we propose two schemes to measure the feature-level similarities from different perspectives. Then, a joint framework of feature selection and clustering is developed to capture the underlying cluster information. The objective function is efficiently optimized by our proposed iterative algorithm. Extensive experimental results on six publicly available databases demonstrate that \(\hbox {U}^2\hbox {F}^2\hbox {S}^2\) outperforms the state-of-the-arts.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Notes

  1. http://www.cad.zju.edu.cn/home/dengcai/Data/MLData.html.

  2. http://www.sheffield.ac.uk/eee/research/iel/research/face.

  3. http://www.cl.cam.ac.uk/research/dtg/attarchive/facedatabase.html.

  4. http://www.cad.zju.edu.cn/home/dengcai/Data/MCFS.html.

  5. http://www.cs.cmu.edu/yiyang/Publications.html.

  6. https://sites.google.com/site/qianmingjie/home/publications.

  7. https://sites.google.com/site/dyhan0920/.

  8. http://www.escience.cn/people/fpnie/papers.html.

References

  1. Wang R, Nie F, Hong R, Chang X, Yang X, Yu W (2017) Fast and orthogonal locality preserving projections for dimensionality reduction. IEEE Trans Image Process 26(10):5019–5030

    Article  MathSciNet  MATH  Google Scholar 

  2. Han D, Kim J (2015) Unsupervised simultaneous orthogonal basis clustering feature selection. In: Proceedings of the IEEE conference on computer vision and pattern recognition. IEEE Press, Boston, pp 5016–5023

  3. He R, Tan T, Wang L, Zheng WS (2012) \(l_{2, 1}\) regularized correntropy for robust feature selection. In: Proceedings of the IEEE conference on computer vision and pattern recognition. IEEE Press, Providence, pp 2504–2511

  4. Wang K, He R, Wang L, Wang W, Tan T (2016) Joint feature selection and subspace learning for cross-modal retrieval. IEEE Trans Pattern Anal Mach Intell 38(10):2010–2023

    Article  Google Scholar 

  5. Nie F, Zhu W, Li X (2016) Unsupervised feature selection with structured graph optimization. In: Thirtieth AAAI conference on artificial intelligence. AAAI Press, Phoenix, AZ, pp 1302–1308

  6. Nie F, Xiang S, Jia Y, Zhang C, Yan S (2008) Trace ratio criterion for feature selection. In: Proceedings of the twenty-third AAAI conference on artificial intelligence. AAAI Press, Chicago, IL, pp 671–676

  7. Yang S, Hou C, Nie F, Wu Y (2012) Unsupervised maximum margin feature selection via \(L_{2, 1}\)-norm minimization. Neural Comput Appl 21:1791–1799

    Article  Google Scholar 

  8. Wang D, Nie F, Huang H (2014) Unsupervised feature selection via unified trace ratio formulation and K-means clustering (TRACK). In: European conference, ECML PKDD, vol 8726. Nancy, France, pp 306–321

  9. He X, Cai D, Niyogi P (2005) Laplacian score for feature selection. In: Conference on neural information processing systems. NIPS, British Columbia, p 189

  10. Cai D, Zhang C, He X (2010) Unsupervised feature selection for multi-cluster data. In: Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining. ACM Press, Washington, pp 333–342

  11. Yang Y, Shen HT, Ma Z, Huang Z, Zhou X (2011) \(l_{2, 1}\)-norm regularized discriminative feature selection for unsupervised learning. In: 21st International joint conference on artificial intelligence. AAAI Press, Barcelona, p 1589

  12. Li Z, Yang Y, Liu J, Zhou X, Lu H (2012) Unsupervised feature selection using nonnegative spectral analysis. In: 26th Conference on artificial intelligence. AAAI Press, Toronto, pp 1026–1032

  13. Qian M, Zhai C (2013) Robust unsupervised feature selection. In: 23rd International joint conference on artificial intelligence. AAAI Press, Beijing, pp 1621–1627

  14. Lee DD, Seung HS (1999) Learning the parts of objects by non-negative matrix factorization. Nature 401:788–791

    Article  MATH  Google Scholar 

  15. Ng AY, Jordan MI, Weiss Y (2001) On spectral clustering: analysis and an algorithm. In: Conference on neural information processing systems. NIPS, British Columbia, pp 849–856

  16. Stella XY, Shi J (2003) Multiclass spectral clustering. In: 9th IEEE international conference on computer vision. IEEE Press, Nice, pp 313–319

  17. Shi J, Malik J (2000) Normalized cuts and image segmentation. IEEE Trans Pattern Anal Mach Intell 22:888–905

    Article  Google Scholar 

  18. Von Luxburg U (2007) A tutorial on spectral clustering. Stat Comput 17:395–416

    Article  MathSciNet  Google Scholar 

  19. Nie F, Ding C, Luo D, Huang H (2010) Improved MinMax cut graph clustering with nonnegative relaxation. In: European conference on machine learning and knowledge discovery in databases, vol 6322. Springer, pp 451–466

  20. Viklands T (2006) Algorithms for the weighted orthogonal procrustes problem and other least squares problems. Doctoral dissertation, Datavetenskap

  21. Lee DD, Seung HS (2001) Algorithms for non-negative matrix factorization. In: Conference on neural information processing systems. NIPS, British Columbia, pp 556–562

  22. Nie F, Huang H, Cai X, Ding CH (2010) Efficient and robust feature selection via joint \(l_{2, 1}\)-norms minimization. In: Conference on neural information processing systems. NIPS, British Columbia, pp 1813–1821

  23. Nene SA, Nayar SK, Murase H (1996) Columbia object image library (COIL-20). Technical report, CUCS-005-96

  24. Fanty MA, Cole RA (1990) Spoken letter recognition. In: Conference on neural information processing systems. NIPS, Denver, pp 220–226

  25. Hull JJ (1994) A database for handwritten text recognition research. IEEE Trans Pattern Anal Mach Intell 16:550–554

    Article  Google Scholar 

  26. Georghiades AS, Belhumeur PN, Kriegman DJ (2001) From few to many: illumination cone models for face recognition under variable lighting and pose. IEEE Trans Pattern Anal Mach Intell 23:643–660

    Article  Google Scholar 

  27. Graham DB, Allinson NM (1998) Characterising virtual eigen signatures for general purpose face recognition. Face Recognit 163:446–456

    Article  Google Scholar 

  28. Samaria FS, Harter AC (1994) Parameterisation of a stochastic model for human face identification. In: 2nd IEEE workshop on applications of computer vision. IEEE Press, Sarasota, pp 138–142

Download references

Acknowledgements

This work is funded by the National Natural Science Foundation of China (Grant No. 61402079), the Foundation for Innovative Research Groups of the NSFC (Grant No. 71421001), the Open Project Program of the National Laboratory of Pattern Recognition (NLPR, No. 201600022).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yanqing Guo.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Zheng, X., Guo, Y., Guo, J. et al. \(\hbox {U}^2\hbox {F}^2\hbox {S}^2\): Uncovering Feature-level Similarities for Unsupervised Feature Selection. Neural Process Lett 49, 1071–1091 (2019). https://doi.org/10.1007/s11063-018-9886-5

Download citation

  • Published:

  • Issue Date:

  • DOI: https://doi.org/10.1007/s11063-018-9886-5

Keywords

Navigation