Abstract
In recent years, with the introduction of manifold learning, graph-based multi-label feature selection has received much attention and achieved state-of-the-art performance. However, improving the graph quality is still a problem that needs to be solved urgently. In addition, existing methods only focus on correlation learning and ignore the feature redundancy problem. To solve these problems, we design an adaptive dynamic graph learning method (ADG), which obtains high-quality dynamic similarity graphs by constraining the Laplacian rank of similarity graphs. Moreover, high-quality dynamic redundancy graphs can also be obtained using ADG, which can better solve the feature redundancy problem. Then, using the \(L_{2,1}\) norm as a sparsity constraint, we propose sparse low-redundant multi-label feature selection (SLADG) with adaptive dynamic dual-graph constraints, and design an efficient scheme to optimize SLADG. Finally, experimental results on eleven publicly available data demonstrate that ADG can obtain high-quality dynamic graphs, and relative to existing state-of-the-art methods, SLADG improves overall performance by an average of at least 3.0427% across the seven evaluated metrics.
Similar content being viewed by others
Explore related subjects
Discover the latest articles, news and stories from top researchers in related subjects.Data Availability
The data that support the findings of this study are openly available in Mulan Library at http://mulan.sourceforge.net/ datasets.html.
References
Chen ZM, Wei XS, Wang P et al (2019) Multi-label image recognition with graph convolutional networks. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition 5177–5186
Chang WC, Jiang D, Yu HF et al (2021) Extreme multi-label learning for semantic matching in product search. Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining 2643–2651
Ye H, Chen Z, Wang DH et al (2020) Pretrained generalized autoregressive model with adaptive probabilistic label clusters for extreme multi-label text classification. International Conference on Machine Learning. PMLR 10809–10819
Huang J, Xu LC, Qian K et al (2021) Multi-label learning with missing and completely unobserved labels. Data Min Knowl Disc 35:1061–1086
Khalid S, Khalil T, Nasreen S (2014) A survey of feature selection and feature extraction techniques in machine learning. 2014 science and information conference. IEEE 372–378
Guyon L, Gunn S, Nikravesh M et al (2008) Feature extraction: foundations and applications. vol 207, Springer
Li JD, Cheng KW, Wang SH et al (2017) Feature selection: A data perspective. ACM computing surveys (CSUR) 50(6):1–45
Sheikhpour R, Sarram MA, Gharaghani S et al (2017) A survey on semi-supervised feature selection methods. Pattern Recogn 64:141–158
Ding CC, Zhao M, Lin J et al (2019) Multi-objective iterative optimization algorithm based optimal wavelet filter selection for multi-fault diagnosis of rolling element bearings. ISA Trans 88:199–215
Labani M, Moradi P, Ahmadizar F et al (2018) A novel multivariate filter method for feature selection in text classification problems[J]. Eng Appl Artif Intell 70:25–37
Yao C, Liu YF, Jiang B et al (2017) Lle score: a new filter-based unsupervised feature selection method based on nonlinear manifold embedding and its application to image recognition. IEEE Trans Image Process 26(11):5257–5269
González J, Ortega J, Damas M et al (2019) A new multi-objective wrapper method for feature selection-accuracy and stability analysis for BCI. Neurocomputing 333:407–418
Jadhav S, He HM, Jenkins K (2018) Information gain directed genetic algorithm wrapper feature selection for credit rating. Appl Soft Comput 69:541–553
Maldonado S, López J (2018) Dealing with high-dimensional class-imbalanced datasets: Embedded feature selection for SVM classification. Appl Soft Comput 67:94–105
Kong YC, Yu TW (2018) A graph-embedded deep feedforward network for disease outcome classification and feature selection using gene expression data. Bioinformatics 34(21):3727–3737
Zhang Y, Huo W, Tang J (2024) Multi-label feature selection via latent representation learning and dynamic graph constraints. Pattern Recogn 151:110411
Zhang YL, Yang Y, Li TR et al (2019) A multitask multiview clustering algorithm in heterogeneous situations based on LLE and LE. Knowl-Based Syst 163:776–786
Cai D, Zhang C, He X (2010) Unsupervised feature selection for multi-cluster data. Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining 333–342
Cai ZL, Zhu W (2018) Multi-label feature selection via feature manifold learning and sparsity regularization. Int J Mach Learn Cybern 9:1321–1334
Hu J, Li YH, Gao WF et al (2020) Robust multi-label feature selection with dual-graph regularization. Knowl-Based Syst 203:106126
Huang R, Wu ZJ (2021) Multi-label feature selection via manifold regularization and dependence maximization. Pattern Recogn 120:108149
Gao WF, Li YH, Hu L (2023) Multilabel Feature Selection With Constrained Latent Structure Shared Term. IEEE transactions on neural networks and learning systems 34(3):1253–1262
Zhang J, Luo ZM, Li CD et al (2019) Manifold regularized discriminative feature selection for multi-label learning. Pattern Recogn 95:136–150
Gretton A, Bousquet O, Smola A et al (2005) Measuring statistical dependence with Hilbert-Schmidt norms. International conference on algorithmic learning theory. Berlin, Heidelberg: Springer Berlin Heidelberg 63–77
Li YH, Hu L, Gao WF (2023) Multi-label feature selection via robust flexible sparse regularization. Pattern Recogn 134:109074
Hu JC, Li YH, Xu GC et al (2022) Dynamic subspace dual-graph regularized multi-label feature selection. Neurocomputing 467:184–196
Zhang Y, Ma YC (2022) Non-negative multi-label feature selection with dynamic graph constraints. Knowl-Based Syst 238:107924
Zhang Y, Ma YC (2023) Sparse multi-label feature selection via dynamic graph manifold regularization. Int J Mach Learn Cybern 14(3):1021–1036
Li YH, Hu L, Gao WF (2023) Robust sparse and low-redundancy multi-label feature selection with dynamic local and global structure preservation. Pattern Recogn 134:109120
Wu YH, Bai JX (2024) Sparse low-redundancy multi-label feature selection with constrained laplacian rank. International Journal of Machine Learning and Cybernetics 1–18
Zhang Y, Ma YC, Yang XF (2022) Multi-label feature selection based on logistic regression and manifold learning. Applied Intelligence1–18
Lin YJ, Hu QH, Liu JH et al (2015) Multi-label feature selection based on max-dependency and min-redundancy. Neurocomputing 168:92–103
Lee J, Kim DW (2013) Feature selection for multi-label classification using multivariate mutual information. Pattern Recogn Lett 34(3):349–357
Hashemi A, Dowlatshahi MB, Nezamabadi-Pour H (2020) MFS-MCDM: Multi-label feature selection using multi-criteria decision making. Knowl-Based Syst 206:106365
Chung FR, Graham FC (1997) Spectral graph theory. American Mathematical Society
Fan K (1949) On a theorem of Weyl concerning eigenvalues of linear transformations I. Proc Natl Acad Sci 35(11):652–655
Han JQ, Sun ZY, Hao HW (2015) Selecting feature subset with sparsity and low redundancy for unsupervised learning. Knowl-Based Syst 86:210–223
Tang C, Liu XW, Li MM et al (2018) Robust unsupervised feature selection via dual self-representation and manifold regularization. Knowl-Based Syst 145:109–120
Huang J, Nie FP, Huang H (2015) A new simplex sparse learning model to measure data similarity for clustering. Twenty-fourth international joint conference on artificial intelligence
Lee DD, Seung HS (1999) Learning the parts of objects by non-negative matrix factorization. Nature 401(6755):788–791
Fan YL, Liu JH, Weng W et al (2021) Multi-label feature selection with constraint regression and adaptive spectral graph. Knowl-Based Syst 212:106621
Tsoumakas G, Spyromitros-Xioufis E, Vilcek J et al (2011) Mulan: A java library for multi-label learning. The Journal of Machine Learning Research 12:2411–2414
Zhang ML, Zhou ZH (2007) ML-KNN: a lazy learning approach to multi-label learning. Pattern Recogn 40(7):2038–2048
Friedman M (1940) A Comparison of Alternative Tests of Significance for the Problem of m Rankings. Ann Math Stat 11(1):86–92
Dunn OJ (1961) Multiple Comparisons among Means. Publ Am Stat Assoc 56(293):52–64
Acknowledgements
This work was supported by National Natural Science Foundation of China key Project: Experimental Study on Lagrangian Turbulence Structure and its Influence on Transport diffusion (11732010), and the Natural Science Foundation of China (12202309).
Author information
Authors and Affiliations
Corresponding author
Ethics declarations
Conflicts of Interest
The authors declared that they have no conflicts of interest to this work. We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.
Additional information
Publisher's Note
Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.
Rights and permissions
Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.
About this article
Cite this article
Wu, Y., Bai, J. Sparse low-redundancy multi-label feature selection with adaptive dynamic dual graph constraints. Appl Intell 55, 228 (2025). https://doi.org/10.1007/s10489-024-06205-3
Accepted:
Published:
DOI: https://doi.org/10.1007/s10489-024-06205-3