ABSTRACT
We live in an era of big data, in which feature selection technology is getting more and more attention. Feature selection technology is one of the important methods to reduce the dimension of data. It can select some useful features for learning tasks. The traditional feature selection methods mainly select the useful features by the scores of the features under a certain standard. However, the performance of these methods are less satisfactory in many cases because they ignore the correlation between features. For this article, we present a new unsupervised method by minimizing the Kullback-Leibler(KL) divergence based on graph matching. Firstly, we extract manifold structures from all features of the original data space by using non-negative Local Linear Embedding(NNLLE). Then, we extract manifold structure of each feature by using non-negative local linear embedding (NNLLE). We assess the importance of every feature by minimizing the KL-divergence between the graphs using all features and weighted linear combination of base graphs on each individual feature. At the same time, a global optimization algorithm based on proximal gradient descent framework is proposed. Experiments show that the proposed method is better than many existing unsupervised methods.
- Stephen Boyd and Lieven Vandenberghe. 2004. Convex Optimization.Google Scholar
- D. Cai, C. Zhang, and X. He. 2010. Unsupervised feature selection for multi-cluster data. In SIGKDD. 333–342.Google Scholar
- Liang Du, Xiaolin Lv, Chaohong Ren, and Yan Chen. 2019. A Filter-based Unsupervised Feature Selection Method via Improved Local Structure Preserving. Proceedings - 2019 5th International Conference on Big Data and Information Analytics, BigDIA 2019, 162 – 169.Google ScholarCross Ref
- Liang Du, Chaohong Ren, Xiaolin Lv, Yan Chen, Peng Zhou, and Zhiguo Hu. 2019. Local graph reconstruction for parameter free unsupervised feature selection. IEEE Access 7(2019), 102921–102930.Google ScholarCross Ref
- Liang Du, Xin Ren, Peng Zhou, and Zhiguo Hu. 2020. Unsupervised Dual Learning for Feature and Instance Selection. IEEE Access 8(2020), 170248–170260.Google ScholarCross Ref
- Liang Du and Yi-Dong Shen. 2013. Joint clustering and feature selection. (2013), 241–252.Google Scholar
- Liang Du and Yi-Dong Shen. 2013. Joint clustering and feature selection. In International conference on web-age information management. 241–252.Google ScholarDigital Library
- Mingyu Fan, Xiaojun Chang, Xiaoqin Zhang, Di Wang, and Liang Du. 2017. Top-k Supervise Feature Selection via ADMM for Integer Programming. In IJCAI. 1646–1653.Google Scholar
- Nannan Gu, Mingyu Fan, Liang Du, and Dongchun Ren. 2015. Efficient sequential feature selection based on adaptive eigenspace model. Neurocomputing 161(2015), 199–209.Google ScholarDigital Library
- Quanquan Gu, Zhenhui Li, and Jiawei Han. 2011. Linear Discriminant Dimensionality Reduction.Google Scholar
- X. He, D. Cai, and P. Niyogi. 2006. Laplacian score for feature selection. Neural information processing systems 18 (2006), 507–514.Google Scholar
- Rongyao Hu, Xiaofeng Zhu, Debo Cheng, Wei He, Yan Yan, Jingkuan Song, and Shichao Zhang. 2017. Graph self-representation method for unsupervised feature selection. Neurocomputing 220(2017), 130–137.Google ScholarCross Ref
- Xuan Li, Yi-Dong Shen, Liang Du, and Chen-Yan Xiong. 2010. Exploiting novelty, coverage and balance for topic-focused multi-document summarization. In Proceedings of the 19th ACM international conference on Information and knowledge management. 1765–1768.Google ScholarDigital Library
- Feiping Nie, Sheng Yang, Rui Zhang, and Xuelong Li. 2018. A General Framework for Auto-Weighted Feature Selection via Global Redundancy Minimization. IEEE Transactions on Image Processing PP, 99 (2018), 1–1.Google Scholar
- Z Rached, F Alajaji, and L.L Campbell. 2004. The Kullback-Leibler divergence rate between Markov sources.TIT 50, 5 (2004), 917–921.Google Scholar
- S.T. Roweis and L.K. Saul. 2000. Nonlinear dimensionality reduction by locally linear embedding. Science 290, 5500 (2000), 2323–2326.Google ScholarCross Ref
- Lei Shi, Liang Du, and Yi-Dong Shen. 2014. Robust Spectral Learning for Unsupervised Feature Selection. In ICDM. 977–982.Google Scholar
- De Wang, Feiping Nie, and Heng Huang. 2015. Feature selection via global redundancy minimization. TKDE 27, 10 (2015), 2743–2755.Google ScholarDigital Library
- Shiping Wang, Witold Pedrycz, Qingxin Zhu, and William Zhu. 2015. Subspace learning for unsupervised feature selection via matrix factorization. Pattern Recognition 48, 1 (2015), 10–19.Google ScholarDigital Library
- L. Wolf and A. Shashua. 2005. Feature selection for unsupervised and supervised inference: The emergence of sparsity in a weight-based approach. JMLR 6(2005), 1855–1887.Google ScholarDigital Library
- Zenglin Xu, Irwin King, MR-T Lyu, and Rong Jin. 2010. Discriminative semi-supervised feature selection via manifold regularization. TNNLS 21, 7 (2010), 1033–1047.Google Scholar
- Chao Yao, Ya-Feng Liu, Bo Jiang, Jungong Han, and Junwei Han. 2017. LLE score: A new filter-based unsupervised feature selection method based on nonlinear manifold embedding and its application to image recognition. TIP 26, 11 (2017), 5257–5269.Google ScholarDigital Library
- Jun Yin, Weiming Zeng, and Lai Wei. 2019. Optimal feature extraction methods for classification methods and their applications to biometric recognition. KBS 99, 1 (2019), 112–122.Google Scholar
- Lihi Zelnik-Manor and Pietro Perona. 2005. Self-tuning spectral clustering. In NIPS. 1601–1608.Google Scholar
- Rui Zhang, Feiping Nie, Yunhai Wang, and Xuelong Li. 2019. Unsupervised Feature Selection via Adaptive Multimeasure Fusion. TNNLS (2019).Google Scholar
- Yan Zhang, Zhao Zhang, Sheng Li, Jie Qin, Guangcan Liu, Meng Wang, and Shuicheng Yan. 2018. Unsupervised Nonnegative Adaptive Feature Extraction for Data Representation. TKDE (2018).Google Scholar
- Hua Zhao, Liang Du, Jianglai Wei, and Yalong Fan. 2020. Local Sensitive Dual Concept Factorization for Unsupervised Feature Selection. IEEE Access 8(2020), 133128–133143.Google ScholarCross Ref
- Zheng Zhao and Huan Liu. 2007. Semi-supervised Feature Selection via Spectral Analysis.. In SDM. 641–646.Google Scholar
- Peng Zhou, Liang Du, Mingyu Fan, and Yi-Dong Shen. 2015. An LLE based Heterogeneous Metric Learning for Cross-media Retrieval. In SDM. 64–72.Google Scholar
- Peng Zhou, Liang Du, Mingyu Fan, and Yi-Dong Shen. 2015. An LLE based heterogeneous metric learning for cross-media retrieval. In Proceedings of the 2015 SIAM international conference on data mining. 64–72.Google ScholarCross Ref
- Peng Zhou, Liang Du, Xuejun Li, Yi-Dong Shen, and Yuhua Qian. 2020. Unsupervised Feature Selection with Adaptive Multiple Graph Learning. Pattern Recognition (2020), 107375.Google Scholar
- Wei Zhou, Chengdong Wu, Yugen Yi, and Guoliang Luo. 2017. Structure Preserving Non-negative Feature Self-Representation for Unsupervised Feature Selection. IEEE Access 5(2017), 8792–8803.Google ScholarCross Ref
Recommendations
Unsupervised feature selection by regularized self-representation
By removing the irrelevant and redundant features, feature selection aims to find a compact representation of the original feature with good generalization ability. With the prevalence of unlabeled data, unsupervised feature selection has shown to be ...
Feature self-representation based hypergraph unsupervised feature selection via low-rank representation
Dimension reduction methods always catch many attentions, because it could effectively solve the curse of dimensionality problem. In this paper, we propose an unsupervised feature selection method which could efficiently select a subset of informative ...
Unsupervised Feature Selection with Feature Clustering
WI-IAT '12: Proceedings of the The 2012 IEEE/WIC/ACM International Joint Conferences on Web Intelligence and Intelligent Agent Technology - Volume 01As an effective technique for dimensionality reduction, feature selection has a broad application in different research areas. In this paper, we present a feature selection method based on a novel feature clustering procedure, which aims at partitioning ...
Comments