Skip to main content

Sparse low-redundancy multi-label feature selection with adaptive dynamic dual graph constraints

  • Published:
Applied Intelligence Aims and scope Submit manuscript

Abstract

In recent years, with the introduction of manifold learning, graph-based multi-label feature selection has received much attention and achieved state-of-the-art performance. However, improving the graph quality is still a problem that needs to be solved urgently. In addition, existing methods only focus on correlation learning and ignore the feature redundancy problem. To solve these problems, we design an adaptive dynamic graph learning method (ADG), which obtains high-quality dynamic similarity graphs by constraining the Laplacian rank of similarity graphs. Moreover, high-quality dynamic redundancy graphs can also be obtained using ADG, which can better solve the feature redundancy problem. Then, using the \(L_{2,1}\) norm as a sparsity constraint, we propose sparse low-redundant multi-label feature selection (SLADG) with adaptive dynamic dual-graph constraints, and design an efficient scheme to optimize SLADG. Finally, experimental results on eleven publicly available data demonstrate that ADG can obtain high-quality dynamic graphs, and relative to existing state-of-the-art methods, SLADG improves overall performance by an average of at least 3.0427% across the seven evaluated metrics.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save

Springer+ Basic
$34.99 /Month
  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime
Subscribe now

Buy Now

Price excludes VAT (USA)
Tax calculation will be finalised during checkout.

Instant access to the full article PDF.

Algorithm 1
Fig. 1
Fig. 2
Fig. 3
Fig. 4
Fig. 5
Fig. 6
Fig. 7

Similar content being viewed by others

Explore related subjects

Discover the latest articles, news and stories from top researchers in related subjects.

Data Availability

The data that support the findings of this study are openly available in Mulan Library at http://mulan.sourceforge.net/ datasets.html.

References

  1. Chen ZM, Wei XS, Wang P et al (2019) Multi-label image recognition with graph convolutional networks. Proceedings of the IEEE/CVF conference on computer vision and pattern recognition 5177–5186

  2. Chang WC, Jiang D, Yu HF et al (2021) Extreme multi-label learning for semantic matching in product search. Proceedings of the 27th ACM SIGKDD Conference on Knowledge Discovery and Data Mining 2643–2651

  3. Ye H, Chen Z, Wang DH et al (2020) Pretrained generalized autoregressive model with adaptive probabilistic label clusters for extreme multi-label text classification. International Conference on Machine Learning. PMLR 10809–10819

  4. Huang J, Xu LC, Qian K et al (2021) Multi-label learning with missing and completely unobserved labels. Data Min Knowl Disc 35:1061–1086

    Article  MathSciNet  MATH  Google Scholar 

  5. Khalid S, Khalil T, Nasreen S (2014) A survey of feature selection and feature extraction techniques in machine learning. 2014 science and information conference. IEEE 372–378

  6. Guyon L, Gunn S, Nikravesh M et al (2008) Feature extraction: foundations and applications. vol 207, Springer

  7. Li JD, Cheng KW, Wang SH et al (2017) Feature selection: A data perspective. ACM computing surveys (CSUR) 50(6):1–45

    Article  MATH  Google Scholar 

  8. Sheikhpour R, Sarram MA, Gharaghani S et al (2017) A survey on semi-supervised feature selection methods. Pattern Recogn 64:141–158

    Article  MATH  Google Scholar 

  9. Ding CC, Zhao M, Lin J et al (2019) Multi-objective iterative optimization algorithm based optimal wavelet filter selection for multi-fault diagnosis of rolling element bearings. ISA Trans 88:199–215

    Article  MATH  Google Scholar 

  10. Labani M, Moradi P, Ahmadizar F et al (2018) A novel multivariate filter method for feature selection in text classification problems[J]. Eng Appl Artif Intell 70:25–37

    Article  MATH  Google Scholar 

  11. Yao C, Liu YF, Jiang B et al (2017) Lle score: a new filter-based unsupervised feature selection method based on nonlinear manifold embedding and its application to image recognition. IEEE Trans Image Process 26(11):5257–5269

    Article  MathSciNet  MATH  Google Scholar 

  12. González J, Ortega J, Damas M et al (2019) A new multi-objective wrapper method for feature selection-accuracy and stability analysis for BCI. Neurocomputing 333:407–418

    Article  MATH  Google Scholar 

  13. Jadhav S, He HM, Jenkins K (2018) Information gain directed genetic algorithm wrapper feature selection for credit rating. Appl Soft Comput 69:541–553

    Article  MATH  Google Scholar 

  14. Maldonado S, López J (2018) Dealing with high-dimensional class-imbalanced datasets: Embedded feature selection for SVM classification. Appl Soft Comput 67:94–105

    Article  MATH  Google Scholar 

  15. Kong YC, Yu TW (2018) A graph-embedded deep feedforward network for disease outcome classification and feature selection using gene expression data. Bioinformatics 34(21):3727–3737

    Article  MATH  Google Scholar 

  16. Zhang Y, Huo W, Tang J (2024) Multi-label feature selection via latent representation learning and dynamic graph constraints. Pattern Recogn 151:110411

    Article  MATH  Google Scholar 

  17. Zhang YL, Yang Y, Li TR et al (2019) A multitask multiview clustering algorithm in heterogeneous situations based on LLE and LE. Knowl-Based Syst 163:776–786

    Article  MATH  Google Scholar 

  18. Cai D, Zhang C, He X (2010) Unsupervised feature selection for multi-cluster data. Proceedings of the 16th ACM SIGKDD international conference on Knowledge discovery and data mining 333–342

  19. Cai ZL, Zhu W (2018) Multi-label feature selection via feature manifold learning and sparsity regularization. Int J Mach Learn Cybern 9:1321–1334

    Article  MATH  Google Scholar 

  20. Hu J, Li YH, Gao WF et al (2020) Robust multi-label feature selection with dual-graph regularization. Knowl-Based Syst 203:106126

    Article  MATH  Google Scholar 

  21. Huang R, Wu ZJ (2021) Multi-label feature selection via manifold regularization and dependence maximization. Pattern Recogn 120:108149

    Article  MATH  Google Scholar 

  22. Gao WF, Li YH, Hu L (2023) Multilabel Feature Selection With Constrained Latent Structure Shared Term. IEEE transactions on neural networks and learning systems 34(3):1253–1262

    Article  MATH  Google Scholar 

  23. Zhang J, Luo ZM, Li CD et al (2019) Manifold regularized discriminative feature selection for multi-label learning. Pattern Recogn 95:136–150

    Article  MATH  Google Scholar 

  24. Gretton A, Bousquet O, Smola A et al (2005) Measuring statistical dependence with Hilbert-Schmidt norms. International conference on algorithmic learning theory. Berlin, Heidelberg: Springer Berlin Heidelberg 63–77

  25. Li YH, Hu L, Gao WF (2023) Multi-label feature selection via robust flexible sparse regularization. Pattern Recogn 134:109074

    Article  MATH  Google Scholar 

  26. Hu JC, Li YH, Xu GC et al (2022) Dynamic subspace dual-graph regularized multi-label feature selection. Neurocomputing 467:184–196

    Article  MATH  Google Scholar 

  27. Zhang Y, Ma YC (2022) Non-negative multi-label feature selection with dynamic graph constraints. Knowl-Based Syst 238:107924

    Article  MATH  Google Scholar 

  28. Zhang Y, Ma YC (2023) Sparse multi-label feature selection via dynamic graph manifold regularization. Int J Mach Learn Cybern 14(3):1021–1036

    Article  MATH  Google Scholar 

  29. Li YH, Hu L, Gao WF (2023) Robust sparse and low-redundancy multi-label feature selection with dynamic local and global structure preservation. Pattern Recogn 134:109120

    Article  MATH  Google Scholar 

  30. Wu YH, Bai JX (2024) Sparse low-redundancy multi-label feature selection with constrained laplacian rank. International Journal of Machine Learning and Cybernetics 1–18

  31. Zhang Y, Ma YC, Yang XF (2022) Multi-label feature selection based on logistic regression and manifold learning. Applied Intelligence1–18

  32. Lin YJ, Hu QH, Liu JH et al (2015) Multi-label feature selection based on max-dependency and min-redundancy. Neurocomputing 168:92–103

    Article  MATH  Google Scholar 

  33. Lee J, Kim DW (2013) Feature selection for multi-label classification using multivariate mutual information. Pattern Recogn Lett 34(3):349–357

    Article  MATH  Google Scholar 

  34. Hashemi A, Dowlatshahi MB, Nezamabadi-Pour H (2020) MFS-MCDM: Multi-label feature selection using multi-criteria decision making. Knowl-Based Syst 206:106365

    Article  MATH  Google Scholar 

  35. Chung FR, Graham FC (1997) Spectral graph theory. American Mathematical Society

    MATH  Google Scholar 

  36. Fan K (1949) On a theorem of Weyl concerning eigenvalues of linear transformations I. Proc Natl Acad Sci 35(11):652–655

    Article  MathSciNet  MATH  Google Scholar 

  37. Han JQ, Sun ZY, Hao HW (2015) Selecting feature subset with sparsity and low redundancy for unsupervised learning. Knowl-Based Syst 86:210–223

    Article  MATH  Google Scholar 

  38. Tang C, Liu XW, Li MM et al (2018) Robust unsupervised feature selection via dual self-representation and manifold regularization. Knowl-Based Syst 145:109–120

    Article  MATH  Google Scholar 

  39. Huang J, Nie FP, Huang H (2015) A new simplex sparse learning model to measure data similarity for clustering. Twenty-fourth international joint conference on artificial intelligence

  40. Lee DD, Seung HS (1999) Learning the parts of objects by non-negative matrix factorization. Nature 401(6755):788–791

  41. Fan YL, Liu JH, Weng W et al (2021) Multi-label feature selection with constraint regression and adaptive spectral graph. Knowl-Based Syst 212:106621

    Article  Google Scholar 

  42. Tsoumakas G, Spyromitros-Xioufis E, Vilcek J et al (2011) Mulan: A java library for multi-label learning. The Journal of Machine Learning Research 12:2411–2414

    MathSciNet  MATH  Google Scholar 

  43. Zhang ML, Zhou ZH (2007) ML-KNN: a lazy learning approach to multi-label learning. Pattern Recogn 40(7):2038–2048

    Article  MATH  Google Scholar 

  44. Friedman M (1940) A Comparison of Alternative Tests of Significance for the Problem of m Rankings. Ann Math Stat 11(1):86–92

    Article  MathSciNet  MATH  Google Scholar 

  45. Dunn OJ (1961) Multiple Comparisons among Means. Publ Am Stat Assoc 56(293):52–64

    Article  MathSciNet  MATH  Google Scholar 

Download references

Acknowledgements

This work was supported by National Natural Science Foundation of China key Project: Experimental Study on Lagrangian Turbulence Structure and its Influence on Transport diffusion (11732010), and the Natural Science Foundation of China (12202309).

Author information

Authors and Affiliations

Authors

Corresponding author

Correspondence to Yanhong Wu.

Ethics declarations

Conflicts of Interest

The authors declared that they have no conflicts of interest to this work. We declare that we do not have any commercial or associative interest that represents a conflict of interest in connection with the work submitted.

Additional information

Publisher's Note

Springer Nature remains neutral with regard to jurisdictional claims in published maps and institutional affiliations.

Rights and permissions

Springer Nature or its licensor (e.g. a society or other partner) holds exclusive rights to this article under a publishing agreement with the author(s) or other rightsholder(s); author self-archiving of the accepted manuscript version of this article is solely governed by the terms of such publishing agreement and applicable law.

Reprints and permissions

About this article

Check for updates. Verify currency and authenticity via CrossMark

Cite this article

Wu, Y., Bai, J. Sparse low-redundancy multi-label feature selection with adaptive dynamic dual graph constraints. Appl Intell 55, 228 (2025). https://doi.org/10.1007/s10489-024-06205-3

Download citation

  • Accepted:

  • Published:

  • DOI: https://doi.org/10.1007/s10489-024-06205-3

Keywords