Abstract
Like traditional single-label learning, multi-label learning is also faced with the problem of dimensional disaster. Feature selection is an effective technique for dimensionality reduction and learning efficiency improvement of high-dimensional data. This paper combined logistic regression, manifold learning, and sparse regularization to construct a joint framework for multi-label feature selection (LMFS). Firstly, the sparsity of the feature weight matrix is constrained by the L2, 1-norm. Secondly, the feature manifold and label manifold can constrain the feature weight matrix to fit the information of data and label better. An iterative updating algorithm is designed, and the convergence of the algorithm is proved. Finally, the LMFS algorithm is compared with DRMFS, SCLS, and other algorithms on eight classical multi-label data sets. The experimental results show the effectiveness of the LMFS algorithm.
Similar content being viewed by others
References
Cai J, Luo JW, Wang SL et al (2018) Feature selection in machine learning: a new perspective[J]. Neurocomputing 300:70–79
Tang C, Liu XW, Zhu XZ et al (2019) Feature selective projection with low-rank embedding and dual laplacian regularization[J]. IEEE Trans Knowl Data Eng 32(9):1747–1760
Bermingham ML, Pong-Wong R, Spiliopoulou A et al (2015) Application of high-dimensional feature selection: evaluation for genomic prediction in man[J]. Scientific Reports 5:10312
Sun X, Liu Y H, Li J et al (2012) Using cooperative game theory to optimize the feature selection problem[J]. Neurocomputing 97:86–93
Zhang R, Nie FP, Li XL et al (2019) Feature selection with multi-view data: a survey[J]. Information Fusion 50:158–167
Ding CC, Zhao M, Lin J et al (2019) Multi-objective iterative optimization algorithm based optimal wavelet filter selection for multi-fault diagnosis of rolling element bearings[J]. ISA Transactions 82:199–215
Labani M, Moradi P, Ahmadizar F et al (2018) A novel multivariate filter method for feature selection in text classification problems[J]. Engineering Applications of Artificial Intelligence 70:25–37
Yao C, Liu YF, Jiang B et al (2017) LLE score: a new filter-based unsupervised feature selection method based on nonlinear manifold embedding and its application to image recognition[J]. IEEE Transactions on Image Processing 26(11):5257–5269
Gonzalez J, Ortega J, Damas M et al (2019) A new multi-objective wrapper method for feature selection - accuracy and stability analysis for BCI[J]. Neurocomputing 333:407–418
Swati J, Hongmei H, Karl J (2018) Information gain directed genetic algorithm wrapper feature selection for credit rating[J]. Applied Soft Computing 69:541–553
Maldonado S, López J (2018) Dealing with high-dimensional class-imbalanced datasets: embedded feature selection for SVM classification[J]. Applied Soft Computing 67:94–105
Kong YC, Yu TW (2018) A graph-embedded deep feedforward network for disease outcome classification and feature selection using gene expression data[J]. Bioinformatics 34(21):3727– 3737
Yang XH, Jiang XY, Tian CX et al (2020) Inverse projection group sparse representation for tumor classification: a low rank variation dictionary approach[J]. Knowledge-Based Systems 196:105768
Deng T Q, Ye D S, Ma R, et al. (2020) Low-rank local tangent space embedding for subspace clustering[J]. Inf Sci 508:1–21
Xiao Q, Dai JH, Luo JW et al (2019) Multi-view manifold regularized learning-based method for prioritizing candidate disease miRNAs[J]. Knowledge-Based Systems 175:118–129
Tang C, Zheng X, Liu XW et al (2021) Cross-view locality preserved diversity and consensus learning for multi-view unsupervised feature selection[J]. IEEE Transactions on Knowledge and Data Engineering 99:1–1
Tang C, Liu XW, Li MM et al (2018) Robust unsupervised feature selection via dual self-representation and manifold regularization[J]. Knowledge-Based Systems 145:109–120
Sun ZQ, Zhang J, Dai L et al (2019) Mutual information based multi-label feature selection via constrained convex optimization[J]. Neurocomputing 329:447–456
Zhang P, Liu GX, Gao WF (2019) Distinguishing two types of labels for multi-label feature selection[J]. Pattern Recogn 95:72–82
Chen LL, Chen DG (2019) Alignment based feature selection for multi-label learning[J]. Neural Processing Letters 50(7):28–36
Chen SB, Zhang YM, Ding CHQ et al (2019) Extended adaptive lasso for multi-class and multi-label feature selection[J]. Knowledge-Based Systems 173:28–36
Zhang J, Luo ZM, Li CD et al (2019) Manifold regularized discriminative feature selection for multi-label learning[J]. Pattern Recognition 95:136–150
Cai ZL, Zhu W (2018) Multi-label feature selection via feature manifold learning and sparsity regularization[J]. Int J Machine Learning Cybern 9(8):1321–1334
Hu JC, Li YH, Gao WF et al (2020) Robust multi-label feature selection with dual-graph regularization[J]. Knowledge-Based Systems 203:106126
Li Q, Xie B, You J et al (2016) Correlated logistic model with elastic net regularization for multilabel image classification[J]. IEEE Transactions on Image Processing 25(8):3801–3813
Sato T, Takano Y, Miyashiro R et al (2016) Feature subset selection for logistic regression via mixed integer optimization[J]. Computational Optimization and Applications 64(3):865– 880
Yang ZY, Liang Y, Zhang H et al (2018) Robust sparse logistic regression with the Lq (0 < q < 1) regularization for feature selection using gene expression data[J]. IEEE Access 6:68586–68595
Pan XL, Xu YT (2021) A safe feature elimination rule for L1-regularized logistic regression[J]. IEEE Transactions on Pattern Analysis and Machine Intelligence. https://doi.org/10.1109/TPAMI.2021.3071138
Zhang R, Nie FP, Li X (2017) Self-weighted supervised discriminative feature selection[J]. IEEE Trans Neural Netw Learn Syst 29(8):3913–3918
Zhang ML, Zhou ZH (2007) ML-KNN: a lazy learning approach to multi-label learning[J]. Pattern Recognition 40(7):2038–2048
Lee J, Kim DW (2017) SCLS: multi-label feature selection based on scalable criterion for large label set[J]. Pattern Recognition 66:342–352
Lin YJ, Hu QH, Liu JH et al (2015) Multi-label feature selection based on max-dependency and min-redundancy[J]. Neurocomputing 168:92–103
Lee J, Lim H, Kim DW (2012) Approximating mutual information for multi-label feature selection[J]. Electron Lett 48(15):929– 930
Lee J, Kim DW (2015) Fast multi-label feature selection based on information-theoretic feature ranking[J]. Pattern Recognition 48(9):2761–2771
Dougherty J, Kohavi R, Sahami M et al (1995) Supervised and unsupervised discretization of continuous features[J]. In: Machine learning: proceedings of the 12th international conference, vol 2, pp 194–202
Demiar J, Schuurmans D (2006) Statistical comparisons of classifiers over multiple data sets[J]. Journal of Machine Learning Research 7(1):1–30
Acknowledgments
This work was supported by the Natural Science Foundation of China (61976130), the Key Research and Development Project of Shaanxi Province (2018KW-021), the Natural Science Foundation of Shaanxi Province (2020JQ-923).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflict of Interests
The authors declared that they have no conflicts of interest to this work. We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.
Additional information
Publisher’s note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
About this article
Cite this article
Zhang, Y., Ma, Y. & Yang, X. Multi-label feature selection based on logistic regression and manifold learning. Appl Intell 52, 9256–9273 (2022). https://doi.org/10.1007/s10489-021-03008-8
Accepted:
Published:
Issue Date:
DOI: https://doi.org/10.1007/s10489-021-03008-8